Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Question] Distributed Training #1

Open
zaptrem opened this issue Nov 5, 2024 · 2 comments
Open

[Question] Distributed Training #1

zaptrem opened this issue Nov 5, 2024 · 2 comments

Comments

@zaptrem
Copy link

zaptrem commented Nov 5, 2024

Does this support distributed training (e.g., DDP/FSDP)? Thanks for sharing!

@ClashLuke
Copy link
Owner

Hey, thank you for your interest!
Everything should support DDP, but it doesn't have any autosharding of optimizer states and computation.

PaLM-SFAdamW might support FSDP, but SOAP won't distribute well.

@LLouice
Copy link

LLouice commented Nov 9, 2024

Hi @ClashLuke, could you please include a section in the README regarding DDP/DeepSpeed/FSDP compatibility? This addition would greatly assist users in applying these techniques in their practical projects!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants