-
Notifications
You must be signed in to change notification settings - Fork 578
Update the rowwise adagrad optimizer to leverage optimizer state offloading, v3 #4133
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
✅ Deploy Preview for pytorch-fbgemm-docs ready!
To edit notification comments on pull requests, go to your Netlify project configuration. |
This pull request was exported from Phabricator. Differential Revision: D74827718 |
This pull request was exported from Phabricator. Differential Revision: D74827718 |
…oading, v3, backend (pytorch#4133) Summary: X-link: facebookresearch/FBGEMM#1214 This diff adds support for leveraging optimizer state offloading to make optimizer state updates, starting with the rowwise adagrad optimizer. - Add compile-time flag `kEnableOptimizerOffloading` to the table update kernel to enable handling optimizer offloading, starting with the rowwise adagrad case - Propagate the compile-time flag upwards to `embedding_backward_split_template.cu`, where it is a runtime user-supplied boolean argument Differential Revision: D74827718
This pull request was exported from Phabricator. Differential Revision: D74827718 |
…oading, v3, backend (pytorch#4133) Summary: Pull Request resolved: pytorch#4133 X-link: facebookresearch/FBGEMM#1214 This diff adds support for leveraging optimizer state offloading to make optimizer state updates, starting with the rowwise adagrad optimizer. - Add compile-time flag `kEnableOptimizerOffloading` to the table update kernel to enable handling optimizer offloading, starting with the rowwise adagrad case - Propagate the compile-time flag upwards to `embedding_backward_split_template.cu`, where it is a runtime user-supplied boolean argument Differential Revision: D74827718
…oading, v3, backend (pytorch#4133) Summary: X-link: facebookresearch/FBGEMM#1214 This diff adds support for leveraging optimizer state offloading to make optimizer state updates, starting with the rowwise adagrad optimizer. - Add compile-time flag `kEnableOptimizerOffloading` to the table update kernel to enable handling optimizer offloading, starting with the rowwise adagrad case - Propagate the compile-time flag upwards to `embedding_backward_split_template.cu`, where it is a runtime user-supplied boolean argument Differential Revision: D74827718
This pull request was exported from Phabricator. Differential Revision: D74827718 |
…oading, v3, backend (pytorch#4133) Summary: X-link: facebookresearch/FBGEMM#1214 This diff adds support for leveraging optimizer state offloading to make optimizer state updates, starting with the rowwise adagrad optimizer. - Add compile-time flag `kEnableOptimizerOffloading` to the table update kernel to enable handling optimizer offloading, starting with the rowwise adagrad case - Propagate the compile-time flag upwards to `embedding_backward_split_template.cu`, where it is a runtime user-supplied boolean argument Differential Revision: D74827718
…oading, v3, backend (pytorch#4133) Summary: X-link: facebookresearch/FBGEMM#1214 This diff adds support for leveraging optimizer state offloading to make optimizer state updates, starting with the rowwise adagrad optimizer. - Add compile-time flag `kEnableOptimizerOffloading` to the table update kernel to enable handling optimizer offloading, starting with the rowwise adagrad case - Propagate the compile-time flag upwards to `embedding_backward_split_template.cu`, where it is a runtime user-supplied boolean argument Differential Revision: D74827718
This pull request was exported from Phabricator. Differential Revision: D74827718 |
…oading, v3, backend (pytorch#4133) Summary: Pull Request resolved: pytorch#4133 X-link: facebookresearch/FBGEMM#1214 This diff adds support for leveraging optimizer state offloading to make optimizer state updates, starting with the rowwise adagrad optimizer. - Add compile-time flag `kEnableOptimizerOffloading` to the table update kernel to enable handling optimizer offloading, starting with the rowwise adagrad case - Propagate the compile-time flag upwards to `embedding_backward_split_template.cu`, where it is a runtime user-supplied boolean argument Differential Revision: D74827718
…oading, v3, backend (pytorch#4133) Summary: Pull Request resolved: pytorch#4133 X-link: facebookresearch/FBGEMM#1214 This diff adds support for leveraging optimizer state offloading to make optimizer state updates, starting with the rowwise adagrad optimizer. - Add compile-time flag `kEnableOptimizerOffloading` to the table update kernel to enable handling optimizer offloading, starting with the rowwise adagrad case - Propagate the compile-time flag upwards to `embedding_backward_split_template.cu`, where it is a runtime user-supplied boolean argument Differential Revision: D74827718
This pull request was exported from Phabricator. Differential Revision: D74827718 |
Summary:
if the optimizer is TBE SSD, and if the user specified it
Differential Revision: D74827718