OpenLLMAI / OpenRLHF

An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & Mixtral)
https://openrlhf.readthedocs.io/
Apache License 2.0
1.72k stars 160 forks source link

fix typos in train_ppo_ray.py #280

Closed mickelliu closed 2 months ago

mickelliu commented 2 months ago

Fix typos. Saw this and was confused when I tried to merge with main locally. Apologize for the repeated PR.

hijkzzz commented 2 months ago

Thanks