-
### 🚀 The feature, motivation and pitch
Please consider adding support for GPTQ and AWQ quantized Mixtral models.
I guess that after #4012 it's technically possible.
### Alternatives
_No r…
-
Title.
https://blog.streamlit.io/ai21_grounded_multi_doc_q-a/
https://arxiv.org/html/2404.09129v1
https://arxiv.org/pdf/2406.02543
https://huggingface.co/papers/2406.02543
https://huggingface…
-
### System Info
PyTorch version: 2.4.1+cu121
Is debug build: False
CUDA used to build PyTorch: 12.1
GPU Type and number: A100 80GB x 1
### Information
- [x] The official example scripts
- [x] M…
-
```
Traceback (most recent call last):
File "/root/miniconda3/lib/python3.10/site-packages/mmengine/runner/_flexible_runner.py", line 1271, in call_hook
getattr(hook, fn_name)(self, **kwargs)…
-
We should be able to support intel GPUs! We are using the intel developer cloud. Please advise.
Distributor ID: Ubuntu
Description: Ubuntu 22.04.4 LTS
Release: 22.04
Codename: …
-
I would like to see some new models added to hugging chat, i will provide my reasoning for each. One is a small model(7B), the other is larger (35B).
# Starling Beta 7B
Link: [HuggingFace Model]…
-
First of all, I apologize if I'm asking nonsense. My doubt arises from the need to edit the same image, not a transformed one. Projects like img2img alternative test (automatic1111 script) have pointe…
-
I am attempting to run the `finetune_onevision.sh` script. I've gotten many things sorted out but I am stumped by the `--pretrain_mm_mlp_adapter` argument.
The default value as provided in the scr…
-
Hello, I load pre-trained llava-llama3 SFT weights and fine-tune using LoRA, but get an error when merging weights:
**scripts:**
Training:
```
deepspeed --master_port=$((RANDOM + 10000)) --inclu…
-
### Information
- [X] Docker
- [ ] The CLI directly
### Tasks
- [X] An officially supported command
- [ ] My own modifications
### Reproduction
I tried lora finetuning a smaller varia…