ymcui / Chinese-LLaMA-Alpaca-2

中文LLaMA-2 & Alpaca-2大模型二期项目 + 64K超长上下文模型 (Chinese LLaMA-2 & Alpaca-2 LLMs with 64K long context models)
Apache License 2.0
7k stars 570 forks source link

关于强化学习 #506

Closed wuhuanon closed 5 months ago

wuhuanon commented 5 months ago

提交前必须检查以下项目

问题类型

其他问题

基础模型

Others

操作系统

Linux

详细描述问题

warmup_rate 参数是干嘛用的,可不可以不用这个参数训练

依赖情况(代码类问题务必提供)

No response

运行日志或截图

No response

GeekDream-x commented 5 months ago

这是设置学习率预热的训练步数占比。使用DeepSpeed-Chat训练代码的话可以参考run_llama2_7b.sh使用参数num_warmup_steps直接指定预热的步数

github-actions[bot] commented 5 months ago

This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your consideration.

github-actions[bot] commented 5 months ago

Closing the issue, since no updates observed. Feel free to re-open if you need any further assistance.