You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
For whatever reason, changing the map_location (an argument to torch.load) inside of _convert(...) in loader.py from cuda to cpu made it work on my 24 GB GPU... and the weights seemed to have ended up on GPU anyway?
That is, the following prints cuda:0 for every parameter even with map_location set to cpu:
I assume there's an implicit .to('cuda') call somewhere in the code, but this still raises the question that if loading the weights to CPU then moving them to GPU works, why do I run out of memory if I try to load them directly to GPU (i.e. with map_location set to cuda)? Does the torch.load call load additional weights that are not actually used?
can you add the option to runn it on a cpu?
The text was updated successfully, but these errors were encountered: