THUDM / ChatGLM-6B

ChatGLM-6B: An Open Bilingual Dialogue Language Model | 开源双语对话语言模型
Apache License 2.0
40.47k stars 5.19k forks source link

[BUG/Help] <max_sequence length> #1348

Open cxjtju opened 1 year ago

cxjtju commented 1 year ago

Is there an existing issue for this?

Current Behavior

基于chatglm-6b进行增量预训练,请问训练数据每个样本的长度不能超过2048个token吗?

Expected Behavior

No response

Steps To Reproduce

长度超过2048会报ValueError: 130004 is not in list 参考这个repohttps://github.com/shibing624/MedicalGPT

Environment

- OS:
- Python:
- Transformers:
- PyTorch:
- CUDA Support (`python -c "import torch; print(torch.cuda.is_available())"`) :

Anything else?

No response