-
Notifications
You must be signed in to change notification settings - Fork 205
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Not Seeing much memory savings with Fp8 optimizer suddenly #1499
Comments
Do you have a snippet to reproduce the issue? Also, what is your PyTorch and torchao version. |
hi @gau-nernst i tried out with torchtitan repo, i just launched the training with Llama 13B and 8B FP8 adamw block_size 128 on H100. i see no memory savings at all: https://github.com/pytorch/torchtitan/blob/main/train_configs/llama3_8b.toml using TP + DP on single node. TP rank 8 DP rank =1. |
I will try to reproduce. Btw, if you switch to AdamW8bit or AdamW4bit, do you observe memory saving? |
@gau-nernst hi what version of ao do you use? |
Latest stable 0.7 |
also which config did you use? did you use bf16 training? |
Default Llama 8B config. All the changes I have mentioned in my previous reply. |
very weird, but i do notice savings on torchtitan but not on my own modified model |
Not Seeing much memory savings with Fp8 optimizer suddenly tried it on Torchtitan Llama 13B
The text was updated successfully, but these errors were encountered: