Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[benchmark] add option to enable CompiledAutograd #1536

Draft
wants to merge 1 commit into
base: main
Choose a base branch
from

Conversation

crcrpar
Copy link
Collaborator

@crcrpar crcrpar commented Dec 10, 2024

What does this PR do?

CompiledAutograd seems to speed up FSDP2 and I checked it with torchtitan.
I however somehow do not find it beneficial for litgpt models.

setting: pjnl-20241209, 8H100

torchtitan Llama-3-8B

This uses activation checkpointing, since the provided config by default uses it -- https://github.com/pytorch/torchtitan/blob/05a8b5e4c1de979c4b49ff36e6b09d6055db29b1/train_configs/llama3_8b.toml#L53-L55

CompiledAutograd Performance (tps) Memory (GB)
N 6244 51.2
Y 7200 43.0

litgpt llama-2-7b-hf

CompiledAutograd Performance (tokens/s/GPU) Memory (GB)
N 11722.76 39.13
Y 10702.33 52.61

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant