-
刚下载的项目
使用官方提供的脚本
hostfile=""
deepspeed --hostfile=$hostfile fine-tune.py \
--report_to "none" \
--data_path "data/belle_chat_ramdon_10k.json" \
--model_name_or_path "baichuan-inc/Ba…
-
### Required prerequisites
- [X] I have read the documentation .
- [X] I have searched the [Issue Tracker](https://github.com/baichuan-inc/baichuan-7B/issues) and [Discussions](https://github.com/bai…
-
**例行检查**
[//]: # (方框内删除已有的空格,填 x 号)
+ [x] 我已确认目前没有类似 issue
+ [x] 我已确认我已升级到最新版本
+ [x] 我已完整查看过项目 README,尤其是常见问题部分
+ [x] 我理解并愿意跟进此 issue,协助测试和提供反馈
+ [x] 我理解并认可上述内容,并理解项目维护者精力有限,**不遵循规则的 issue 可能…
-
官方代码测试:
(python3.8) [baichuan@localhost baichuan-7B]$ python3 generate.py
The model weights are not tied. Please use the `tie_weights` method before using the `infer_auto_device` function.
登鹳雀楼-…
-
尝试autoeval , 评估chatglm2-6b,baichuan-7b 相同的错误如下:
1.chatglm2-6b:
python3 autoeval.py --model chatglm2-6b --lora_path ../../models/chatglm2-6b --eval_data all --device cuda:0
cuda:0
Lo…
-
## model info
* base-model : baichuan-7b
* base-context-size : 4096
Did this phenomenon oberserved in your experiments?
In short context-window: Ntk > Yarn
![image](https://github.com/jquesn…
-
### Required prerequisites
- [X] I have read the documentation .
- [X] I have searched the [Issue Tracker](https://github.com/baichuan-inc/baichuan-7B/issues) and [Discussions](https://github.com/bai…
-
Baichuan 2 支持昇腾 NPU 推理,推理所需的modeling、README、示例脚本已发布:Baichuan2-7B、Baichuan2-13B中的Baichuan2-13B的gitee链接失效。请问可以去哪里找相关资料。
-
### Required prerequisites
- [X] I have read the documentation .
- [X] I have searched the [Issue Tracker](https://github.com/baichuan-inc/baichuan-7B/issues) and [Discussions](https://github.com/bai…
-
评估使用的代码:https://github.com/baichuan-inc/Baichuan-7B/blob/main/evaluation/evaluate_mmlu.py
用bf16精度测试 llama2-13-hf 和 baichuan2-13b-base
llama2-13-hf: 0.550
baichuan2-13b-base: 0.564
改了一行代码,用fp32…