Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[REQUEST] Extend offload_states to support models with cpu-based optimizer #6596

Open
kfertakis opened this issue Oct 1, 2024 · 0 comments
Labels
enhancement New feature or request

Comments

@kfertakis
Copy link

Is your feature request related to a problem? Please describe.
The issue is related to #5620 and #6011. The new offload_states API works only with FusedAdam GPU optimizer. Currently there is no way to offload a trainable model that is using a CPU-based optimizer like DeepSpeedCPUAdam.

Describe the solution you'd like
Extend #6011 to support offloading of a model configured with CPU-based DeepSpeedCPUAdam optimizer.

Thanks,

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

1 participant