shibing624 / MedicalGPT

MedicalGPT: Training Your Own Medical GPT Model with ChatGPT Training Pipeline. 训练医疗大模型,实现了包括增量预训练(PT)、有监督微调(SFT)、RLHF、DPO、ORPO。
Apache License 2.0
3.21k stars 488 forks source link

使用deepspeed run run_rm.sh #73

Closed yangzhipeng1108 closed 1 year ago

yangzhipeng1108 commented 1 year ago

Could not estimate the number of tokens of the input, floating-point operations will not be computed Traceback (most recent call last): File "/root/nas-share/chat/MedicalGPT-main/reward_modeling.py", line 643, in main() File "/root/nas-share/chat/MedicalGPT-main/reward_modeling.py", line 615, in main train_result = trainer.train(resume_from_checkpoint=checkpoint) File "/usr/local/conda/lib/python3.9/site-packages/transformers/trainer.py", line 1537, in train return inner_training_loop( File "/usr/local/conda/lib/python3.9/site-packages/transformers/trainer.py", line 1881, in _inner_training_loop self.optimizer.step() AttributeError: 'DummyOptim' object has no attribute 'step'

RewardTrainer是否支持deepspeed

xxm1668 commented 1 year ago

deepspeed版本号问题

yangzhipeng1108 commented 1 year ago

请问大神,什么版本比较合适

xxm1668 commented 1 year ago

0.9.2

yinheng-gh commented 1 year ago

我是在运行pretrain的时候遇到这个问题,把deepspeed改成0.9.2会在accelerate检查依赖版本时提示最低需要0.9.3,然后改成0.9.3,还是会遇到上面那个问题