We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
gpt_options=" \ --experiment-name finetune-$MODEL_TYPE \ --model-parallel-size ${MP_SIZE} \ --mode finetune \ --train-iters 100 \ --resume-dataloader \ $MODEL_ARGS \ --train-data ${train_data} \ --valid-data ${eval_data} \ --distributed-backend nccl \ --lr-decay-style cosine \ --warmup .02 \ --checkpoint-activations \ --save-interval 1200 \ --eval-interval 10000 \ --save "./checkpoints" \ --split 1 \ --eval-iters 10 \ --eval-batch-size 1 \ --zero-stage 2 \ --lr 0.0001 \ --batch-size 1 \ --skip-init \ --fp16 \ --use_lora "
如上,这是在stage2训练,使用finetune_visualglm.py和lora微调,其他程序未改动。但是结果得到的权重达到25G,远超官方公布的权重。想请教下,这是啥情况呢?不使用stage2而使用stage1训练时也才14.5G左右。
The text was updated successfully, but these errors were encountered:
No branches or pull requests
如上,这是在stage2训练,使用finetune_visualglm.py和lora微调,其他程序未改动。但是结果得到的权重达到25G,远超官方公布的权重。想请教下,这是啥情况呢?不使用stage2而使用stage1训练时也才14.5G左右。
The text was updated successfully, but these errors were encountered: