-
Notifications
You must be signed in to change notification settings - Fork 271
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
converging.md #831
Comments
The set of jobs use 8 to 64 GPUs, the breakdown specified in the table.
FSDP 8 means Fully Sharded Data Parallel degree is 8 |
Thanks, so FSDP 8 means dividing params, gradients and optimizer states into 8 parts ? Is FSDP 8, TP 2, CP 2, PP 2 different from TP 2 , FSDP 8, , PP 2, CP 2 ? |
Yeah
There' no difference for the order when we call them (as well as the config order in CLI) |
Thanks again. In case of FSDP 8, TP 2, CP 2, PP 2 , what is the specific order ? Is there a way to trace the collectives of the 4 dimensions ? |
There are two concepts of "order":
You can look at the profiler trace (with tools like perfetto), after dumping them via https://github.com/pytorch/torchtitan/blob/main/torchtitan/config_manager.py#L92 |
Thanks - ["pp", "dp_replicate", "dp_shard", "cp", "tp"], is the order. Is dp_replicate just DDP ? I guess pp, dp_shard, cp and tp can co-exist. Not sure what dp_replicate can co-exist with. |
When not used with "dp_shard" ( |
In the page . Can someone please clarify the the following.
The text was updated successfully, but these errors were encountered: