-
我在使用vllm0.5.0测试chatglm-6b的性能,一开始运行vllm时出现“AttributeError: ‘ChatGLMTokenizer‘ object has no attribute ‘tokenizer‘”的错误,把chatglm-6b中的tokenization_chatglm.py替换了之后运行vllm的benchmark_throughput.py时又出现如下错误,请问怎…
-
### System Info
12600k
64G
RTX 4070 12G
### Who can help?
_No response_
### Information
- [x] The official example scripts
- [ ] My own modified scripts
### Tasks
- [x] An officially supported …
-
# 破解道德限制
![image](https://user-images.githubusercontent.com/10473170/227781289-e4943db9-0647-4569-ade3-7f215efeed74.png)
ypwhs updated
1 month ago
-
-
### Feature request
Support https://huggingface.co/THUDM/chatglm-6b (and its int4 variants) in the Transformers library instead of relying on remote code execution.
### Motivation
This model pe…
-
### Checked other resources
- [X] I searched the Codefuse documentation with the integrated search.
- [X] I used the GitHub search to find a similar question and didn't find it.
- [X] I am sure that …
-
As a user of the Bloop I would like to request the ability to customize the default GPT-2 model with a custom model of my choice for conversation taking.
Currently, Bloop uses the GPT-2 model by …
-
Hi,可以分享一下你的 model 下的目录文件层级具体是怎样吗?如果可以直接 `tree -L 2 model/LexiLaw` 看一下开发者本地文件层级是怎样就好了,谢谢!
我目前是这样:
```
LexiLaw# tree -L 2 model/LexiLaw
model/LexiLaw
├── LexiLaw_finetune
│ ├── pytorch_model-0…
-
### 这是一个?
现有功能优化
### 详细描述
希望能支持chatglmAPI或者是本地部署的chatglm-6b模型api
-
### 错误描述
机器: h100-80g-hbm3
基于以下的chatglm-6b-xxx.json配置测试,在tp1, bs24, inputlen1024下报OOM
![image](https://github.com/user-attachments/assets/cd7eb052-342f-4eb1-ae92-70b55f0a89a6)
![image](https://g…