-
### System Info
```shell
optimum 1.21.4
optimum-habana 1.14.0.dev0
transformers 4.45.2
+------------------------------------------------------------------…
-
Hi @PanQiWei
I'd be most grateful if you could give me a bit of help.
I have been trying to quantize BLOOMZ 175B but can't currently get it done. BLOOMZ has 70 layers, and is a total of 360GB.…
-
Thank you again for your excellent work. I have trained a model mT0 using my own dataset, and it performs well. Now, I am attempting to train bloomz model, but I'm encountering an issue where the trai…
-
# 模型参数支持专区
大家好,PaddleNLP 团队在这里为大家整理了各个模型参数的详细信息,方便大家使用。
## 模型参数
### Base Models
| Model | 0.5B | 1~2B | 3~4B | 6~8B | 13~14B | 30~32B | 50~60B | 65~72B | 110B | >110B |
|:---------:|:--…
-
The article at https://huggingface.co/blog/generative-ai-models-on-intel-cpu mentions that smoothquant was applied on bloomz 7b1 model also. But in https://huggingface.co/mit-han-lab/smoothquant-scale…
-
Operaring system: Ubuntu 22.04.2, Python 3.10.6, CTranslate2 3.16.
When exporting the bigscience/bloomz model using:
_ct2-transformers-converter --force --model bigscience/bloomz --output_dir …
-
Hi thank you very much for this great library! I am really excited about it!
I tried to fine tune bloomz 7B with the 4 bit lora on alpaca by running: python qlora.py --model_name_path my_bloomz_path.…
-
### 🚀 The feature, motivation and pitch
I have fine-tuned multiple LoRA models to act as expert layers within an MOE architecture. How can I leverage a VLLM to accelerate this? Currently, VLLM acce…
-
I have a q4_0 quant of bloomz-176B (created with bloomz.cpp). As with bloomz.cpp, inference on this model is broken.
For example, this command:
llm infer -a bloom -m bloomz_q4_0.bin -p "A short stor…
-
Hello!
Thanks a lot for your job!
I want to finetune bloomz-mt by your Megatron-DeepSpeed,but I can not find a universal version checkpoint of bloomz-mt or bloomz. I only found the bloom universal c…