issues
search
LlamaFamily
/
Llama-Chinese
Llama中文社区,Llama3在线体验和微调模型已开放,实时汇总最新Llama3学习资料,已将所有代码更新适配Llama3,构建最好的中文Llama大模型,完全开源可商用
https://llama.family
14.01k
stars
1.26k
forks
source link
issues
Newest
Newest
Most commented
Recently updated
Oldest
Least commented
Least recently updated
单卡运行正常,双卡报错 `device-side assert triggered`
#262
verymiao
opened
12 months ago
5
为什么部署到dify上会这样提示啊
#261
ghkl98
opened
12 months ago
2
fixed: bitsandbytes exception on windows platform
#260
iisquare
closed
10 months ago
0
OSError: Error no file named pytorch_model.bin, tf_model.h5, model.ckpt.index or flax_model.msgpack found in directory FlagAlpha/Atom-7B.
#259
ZhenshengWu
closed
1 year ago
1
请问 FlagAlpha/Atom-7B-Chat 与 FlagAlpha/Llama2-Chinese-7b-Chat 之间是什么关系?有什么区别?
#258
yyqqing
opened
1 year ago
0
使用lora脚本微调后保存的模型文件如何再次对他进行微调
#257
A1pacin0
opened
1 year ago
2
RuntimeError: handle_0 INTERNAL ASSERT FAILED at "../c10/cuda/driver_api.cpp":15, please report a bug to PyTorch.
#256
hellangleZ
opened
1 year ago
2
如何 finetune 出想要的結果 ?
#255
arthurwu4work
opened
1 year ago
0
请问在哪里可以获取预训练阶段的train_sft_sharegpt.csv文件呢
#254
xyliu-uir
opened
1 year ago
0
还是别用百度网盘了吧,非会员下载太慢。
#253
taozhiyuai
opened
1 year ago
1
微调 prompt
#252
beimingyouyu1
opened
1 year ago
0
docker build -f docker/Dockerfile failed, /tmp/Python-3.10.12 exec pip failed
#251
sgkokocool1
opened
1 year ago
0
双机deepspeed 跑finetune_lora.sh 这个脚本报错RuntimeError: ProcessGroupNCCL is only supported with GPUs, no GPUs found!
#250
Dagoli
opened
1 year ago
0
pretrain脚本运行时,卡在训练读条
#249
Dagoli
opened
1 year ago
9
API部署出错
#248
venxzw
opened
1 year ago
0
关于用vllm加速baichuan2的问题
#247
venxzw
opened
1 year ago
0
请问有好的把长文本数据切分成对话数据的方式吗?
#246
Junglesl
opened
1 year ago
0
Atom-7B benchmark 指标
#245
wanxinzzz
opened
1 year ago
0
希望能提供 gguf 格式模型
#244
yyqqing
opened
1 year ago
0
使用lora微调,造成显存爆炸
#243
FLoutione
opened
1 year ago
1
微调过程中:RuntimeError: element 0 of tensors does not require grad and does not have a grad_fn
#242
Dagoli
opened
1 year ago
3
peft中的其他微调方法,例如ptuning tuning等微调方法没有merge_and_unload(),该如何合并peft的ptuning tuning与预训练模型呢?
#241
xiaobai52HZ
opened
1 year ago
0
fastertransform_backend 编译docker的时候报错
#240
callmezhangchenchenokay
opened
1 year ago
0
请教一下在微调脚本中,怎么设置成多机多卡的分布式训练呢?如果是在k8s集群当中,是怎么启动多机多卡呢?
#239
wyx22210720244
closed
1 year ago
1
请教一下关于做具体的下游任务的问题
#238
HypherX
opened
1 year ago
0
中文词表
#237
NivinaNull
opened
1 year ago
1
跑 Atom-7B-Chat 模型,代码输出不换行
#236
tanghui315
opened
1 year ago
0
对话模型的demo
#235
sssssshf
opened
1 year ago
0
运行Llama2-70B-Chat需要多大显存
#234
AnitaSherry
opened
1 year ago
1
有无计划预训练一下mistral-7b
#233
longgui0318
closed
1 year ago
0
train/pretrain/pretrain.sh
#232
SWT-AITeam
opened
1 year ago
0
merge_muilt_peft_adapter.py使用场景
#231
ai408
opened
1 year ago
0
KeyError: 'llama'
#230
Barry1915
opened
1 year ago
0
accelerate_server.py 运行问题
#229
dybtom
opened
1 year ago
0
lora训练,没效果,eval_accuracy 几乎不变?
#228
wusanshou2017
closed
1 year ago
0
是否考虑训练一个context更长的模型
#227
yuyijiong
closed
1 year ago
0
是否考虑训练一个context更长的模型
#226
yuyijiong
closed
1 year ago
0
是否考虑训练一个context更长的模型
#225
yuyijiong
opened
1 year ago
0
这是否是transformer的版本问题
#224
GluttonousCat
opened
1 year ago
1
RuntimeError: CUDA error: invalid device ordinal Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions.
#223
yfangZhang
opened
1 year ago
0
文本生成场景适用吗
#222
jiaozhusos
opened
1 year ago
0
全量微调13B的模型,大概需要多大的显存
#221
kunden0612
closed
1 year ago
1
中文数据集测试
#220
shuliu586
opened
1 year ago
1
使用lora方式,单机多卡微调是需要怎么怎么修改脚本?
#219
yfangZhang
opened
1 year ago
1
关于lora微调脚本的数据集的问题
#218
tianjianmeng
opened
1 year ago
2
尝试用Chat-7B进行finetune后,lora生成的结果是中文加乱码问题
#217
DJinsis
opened
1 year ago
0
能否提供FlagAlpha/Llama2-Chinese-7b-Chat-4bit模型或者有没有脚本可以对7b进行4bit量化
#216
futureflsl
opened
1 year ago
0
SFT数据完整资料?
#215
gameveloster
opened
1 year ago
0
Llama-2-70b-hf 大佬也能上传一下云盘嘛?谢谢🙏
#214
confucianism72
closed
1 year ago
1
多个用户同时访问gradio出现乱码情况(使用的是llama2-chinese-13b-4bit)
#213
Dragon-gif
opened
1 year ago
0
Previous
Next