issues
search
OpenLLMAI
/
OpenRLHF
An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & Mixtral)
https://openrlhf.readthedocs.io/
Apache License 2.0
1.71k
stars
160
forks
source link
mixtral 8*7B的最佳训练参数,推理参数可以提供一个吗?
#319
Closed
harvinyou
closed
3 weeks ago
hijkzzz
commented
3 weeks ago
大致参考 llama 70b 即可,细微的性能参数要手工调整了。
大致参考 llama 70b 即可,细微的性能参数要手工调整了。