-
您好,我尝试了您提供的chatglm纠错模型 (https://huggingface.co/shibing624/chatglm3-6b-csc-chinese-lora) 效果很好,但没法纠错一些特定领域(例如医疗、金融)内文本。
想请问一下,如果我想补充一些领域内的的纠错语料,是直接在chatglm3-6b-csc-chinese-lora上进行微调吗?还是需要加上您提供的数据集在ch…
-
### System Info
- CPU: INTEL RPL
- GPU Name: NVIDIA GTX 4090
- TensorRT-LLM: tensorrt_llm==0.11.0.dev2024060400
- Container Used: Yes and reproduced in Conda as well
- Driver Version: 555.42.02
…
-
bash ./run_benchmark.sh -m chatglm3-6b -d nf4 -s 1 -bs 1 -in 256 -out 256 -i 1
bash ./run_benchmark.sh -m chatglm3-6b -d int4 -s 1 -bs 1 -in 256 -out 256 -i 1
bash ./run_benchmark.sh -m chatglm3-6b …
-
现在新版本xtuner增加了dispatch后,不支持chatglm3-6b的微调了吗
File "/mnt/afs/xtuner/xtuner/model/sft.py", line 93, in __init__
dispatch_modules(self.llm, use_varlen_attn=use_varlen_attn)
File "/mnt/afs/xtuner/…
-
**Before:**
![e99fd6f8bd6b3cc802c39f03a0adad1](https://github.com/NEFUJing/LawyerLLM/assets/106534091/0a4f119c-b24c-4e8b-a19b-7694784a215a)
**After:**
![504b032158f719b58ddc0b978a906fc](https://git…
-
您好,我使用fastchat进行加载chatglm3-6b模型,
step1 `python3 -m fastchat.serve.controller`
step2 `python3 -m fastchat.serve.model_worker --model-path /ldata/llms/chatglm3-6b`
step3 `python3 -m fastchat.serve.op…
-
### System Info
platform == ubuntu 22.04
transformers == 4.32.2
python == 3.10.12
### Who can help?
_No response_
### Information
- [ ] The official example scripts
- [ ] My own modified s…
-
### System Info
- CPU: X86
- GPU: NVIDIA L20
- python
- tensorrt 10.3.0
- tensorrt-cu12 10.3.0
- tensorrt-cu12-bindings 10.3.0
- tensorrt-cu12-libs 10…
-
Hi master Kijai, did your wrapper support this Kolors gguf model.
https://huggingface.co/twodgirl/Kolors-unet-gguf
-
### I tested vllm benchmark_throughput.py and finded that the performance with chunked-prefill-enabled is lower than default, how can I deal this problem
_No response_
### Your current environ…