-
baichuan-7B is an open-source large-scale pre-trained model developed by Baichuan Intelligent Technology. Based on the Transformer architecture, it is a model with 7 billion parameters trained on appr…
-
So my P40 is only using about 70W while generating responses, its not limited in any way(IE. Power delivery or temp)
-
`python generate.py`,输出如下:
```
===================================BUG REPORT===================================
Welcome to bitsandbytes. For bug reports, please submit your error trace to: https://…
-
Traceback (most recent call last):
File "/home/Bloom-Lora/processor/processing.py", line 108, in
instruction_dataset = instruction_dataset.map(group_text,
File "/usr/local/python3/lib/pyth…
-
I am trying to understand the best way to setup prompts and the library for an interactive chat session. It looks like based on the InteractiveModeExecute.cs example, the "bob" personality is only de…
-
### Describe the bug
No matter what model I load, it always produces an error (wizardLM-7B-GPTQ-4bit-128g, wizard-vicuna-7b-uncensored-gptq-4bit-128g no-act-order safetensors).
### Is there an exist…
-
I was trying to do an apples-to-apple shootout on GPTQ vs the new llama.cpp k-quants (memory usage, speed, etc) but ran into a bump with perplexity. It looks like exllama loads a jsonl formatted versi…
-
[[BELLE](https://github.com/LianjiaTech/BELLE)](https://huggingface.co/datasets/BelleGroup/multiturn_chat_0.8M)
这里公开了 8M的 80W条 多轮对话的语料
格式是这样:
```
instruction: 指令(这里是Human/Assistant,多轮对话上下文)
inp…
-
Traceback (most recent call last):
File "/data/Chinese-Vicuna/tools/convert_pth_to_ggml.py", line 64, in
with open(fname_hparams, "r") as f:
FileNotFoundError: [Errno 2] No such file or directo…
-
本人使用文档提供的[colab链接](https://colab.research.google.com/drive/1OLCJ-ZHogm5O3RdyUDY83YfgnCXdHNXp?usp=sharing), 发现不能运行在colab上
本人未有修改过提供的脚本, 也未有改变脚本参数和命令参数,然后其他参数、数据都和你们一致。
运行了如下三个命令
```
!git clone …