Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

get_train_val_test_loader_from_train does not produce consistent splits across machines #43

Open
crrrr30 opened this issue Jul 3, 2024 · 1 comment

Comments

@crrrr30
Copy link

crrrr30 commented Jul 3, 2024

First off, thanks for sharing your complete code, which has been more than helpful for my current research.

Observation: It was rather challenging to have identified this issue. In particular, I noted that evaluating on a different machine than training produces unrealistically high Dice and HD95 metrics. Evaluation results, however, remain reproducible on the same machine.

Problem: According to this post, the return of glob.glob is arbitrary. Indeed, I could confirm that the return of glob.glob in line 247 is different on different machine, although on the same machine it seems to consistently reproduce the same ordering.

Could you share your data split in the paper so that we can reproduce your results? Thanks in advance for your time and attention.

@crrrr30
Copy link
Author

crrrr30 commented Jul 3, 2024

I'll add that it seems to be dependent on the file system. If the machine access a shared directory containing the data, then the training split is reproducible. Otherwise, if different machines have different copies of data (i.e., potentially different metadata), they do not in general output the same glob.glob() results.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant