Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Performance discrepancy when training from scratch w/ PyTorch 1.4 #5

Open
bertinetto opened this issue Feb 20, 2020 · 2 comments
Open

Comments

@bertinetto
Copy link

bertinetto commented Feb 20, 2020

Hi,
Thanks for the nice work and for sharing the code.

I have tried replicating the results of the paper (training from scratch) but with no luck.
I have followed the instructions of the readme and tried both pytorch/cuda-toolkit 1.4/10.0.
For ResNet-10 and ResNet-18, on miniImageNet I am getting a discrepancy between 2% and 3% (absolute).

Thanks in advance for the help!

@bertinetto bertinetto changed the title Performance discrepancy when re-training Performance discrepancy when training from scratch (not using the provided models) Feb 20, 2020
@bertinetto
Copy link
Author

Hi,
For some reason, the problem only occurred when using PyTorch 1.4.
Using PyTorch 1.0 results are reproduced.
Will post here when/if I can find a fix for PyTorch 1.4

@bertinetto bertinetto changed the title Performance discrepancy when training from scratch (not using the provided models) Performance discrepancy when training from scratch w/ PyTorch 1.4 Feb 25, 2020
@mileyan
Copy link
Owner

mileyan commented Feb 25, 2020

Thanks for pointing it out. I will look at the difference between 1.4 and 1.0.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants