We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
想问一下这里是如何实现ema模型的保存呢: 从这里没有看到model被显示的修改,是deepspeed内完成的么?
我有check保存下来的model和ema_model,发现参数是完全一样的,所以不确定是不是使用方式有问题,望解答,感谢
The text was updated successfully, but these errors were encountered:
是sat内部实现的optimizer进行的操作:
https://github.com/THUDM/SwissArmyTransformer/blob/main/sat/ops/fused_ema_adam.py
直接将ema给fuse到optmizer更新的过程中了。
应该不会有问题,之前CogVLM训练的时候能观察到ema权重评测结果高一些。
Sorry, something went wrong.
No branches or pull requests
想问一下这里是如何实现ema模型的保存呢:
从这里没有看到model被显示的修改,是deepspeed内完成的么?
我有check保存下来的model和ema_model,发现参数是完全一样的,所以不确定是不是使用方式有问题,望解答,感谢
The text was updated successfully, but these errors were encountered: