-
Unsloth: Merging 4bit and LoRA weights to 4bit...
This might take 5 minutes...
/usr/local/lib/python3.10/dist-packages/peft/tuners/lora/bnb.py:336: UserWarning: Merge lora module to 4-bit linear m…
-
Hi,
Thanks for the effort of publishing this significant code base etc.!
In Appendix B1 you write:
> In addition to the main Llama 2 model used for evaluation, we also release HarmBench with …
-
[Mistral-7b-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1), [Zephyr-7b-alpha](https://huggingface.co/HuggingFaceH4/zephyr-7b-alpha)
- Mistral-7b outperformed Llama2-13b-hf and gpt-3.5-tur…
-
# Quantization Impact on Model Accuracy | Slightwind
Mistral-7B’s performance on 5-shot MMLU 如果对测试细节不感兴趣,只需要看下面给出的汇总表格即可。
Overview 量化/非量化版本的 Mistral-7B-v0.1 模型在 5-shot MMLU 上的表现:
Quant Type Compute D…
-
`r2ai -mm` is expected (according to doc `-h`) to list all downloaded models:
```
[r2ai:0x000061d0]> -mm
??? TheBloke/Mistral-7B-Instruct-v0.1-GGUF
2.96G cognitivecomputations/dolphin-…
-
# Prerequisites
Please answer the following questions for yourself before submitting an issue.
- [x ] I am running the latest code. Development is very rapid so there are no tagged versions as o…
-
Is there a way to output the summary statistics of the run on stdout? For example, if I run mistral_7b.yaml, how can I output the statistics on stdout?
-
**Describe**
Thank you for your team's contribution! I would like to fine-tune E5-mistral-7b-instruct for tasks that interest me. Do you have plans to open-source training code? Alternatively, are th…
-
### Issue Description
when running server llava1.6-mistral - unable to load model
nexa server llava1.6-vicuna
objc[12203]: Class GGMLMetalClass is implemented in both /Applications/Nexa.app/Content…
-
Nice work.
I am trying to use fastchat to train a mistral model. however, I wonder why the following code is hard code for only vicuna.
[https://github.com/lm-sys/FastChat/blob/main/fastchat/train/…