Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

PyTorch Lightning example #2

Open
tchaton opened this issue Mar 9, 2022 · 1 comment
Open

PyTorch Lightning example #2

tchaton opened this issue Mar 9, 2022 · 1 comment

Comments

@tchaton
Copy link

tchaton commented Mar 9, 2022

Dear team behind mup,

This is some great work! I believe providing a PyTorch Lightning example could help users adopt this library.

I even wonder if this technique could be embedded in an even less boilerplate way. I was thinking about an extension to Pytorch Lightning Tuner which would automatically apply mup and apply the µTransferable Hyperparameters.

I wondered if someone from the mup Team would be interested to investigate those ideas to democratize even further this work.

Best,
T.C

@edwardjhu
Copy link
Collaborator

Hi tchaton,

Thanks for the pointer to the Lightning Tuner. We are not familiar with its usage, but from the page you linked, it looks like one can pass a model to, for example, lr_find along with a grid and the Tuner performs the necessary for loop(s) and returns the best HPs. In other words, one should be able to pass the proxy model, parametrized in muP, to the Tuner and take advantage of both right away.

Perhaps you are thinking about adding an option such as lr_find(model, mup=True, ...) to the Tuner API. The main obstacle is that we still need to let muP know which dimensions go to infinity in the limit by instantiating models of different widths. We also need the user to manually switch optimizers as well. Both are hard to hide inside a Tuner fn call.

Please let us know if you have ideas on how we can make this integration more seamless!

zygi added a commit to flanlabs/mup that referenced this issue Dec 6, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants