-
# Todo
https://github.com/OCA/maintainer-tools/wiki/Migration-to-version-15.0
# Modules to migrate
- [x] account_move_line_product - By @JuanyDForgeflow - #1428
- [ ] product_route_profile -…
-
v 0.6.1
```bash
python quantize.py --model_dir ./hg_weight_3999/ --dtype float16 --qformat int4_awq --export_path ./quantized_int4-awq --calib_size 32
```
```log
Using pad_token, but it is not se…
-
Hi,
I have finetuned Qwen2-VL using Llama-Factory.
I successfully quantized the fine-tuned model as given
```
from transformers import Qwen2VLProcessor
from auto_gptq import BaseQuantizeC…
-
Hi @Qubitium . Since the CPU path is already in gptqmodel, when do you plan to replace auto_gptq to gptqmodel in HuggingFace/optimum? I think we can start an issue in Optimum to let the maintainer kno…
-
### Your current environment
Hello everyone, I need some help here, please. I tried to quantize the JAIS model using GPTQ. Here is my code:
```
from auto_gptq.modeling._base import BaseGPTQForC…
-
### System Info
```shell
Name: optimum
Version: 1.18.0.dev0
Name: transformers
Version: 4.36.0
Name: auto-gptq
Version: 0.6.0.dev0+cu118
CUDA Version: 11.8
Python 3.8.17
```
### Who can help…
-
# Todo
https://github.com/OCA/maintainer-tools/wiki/Migration-to-version-16.0
# Modules to migrate
- [ ] delivery_procurement_group_carrier - By @rousseldenis - #1158
- [x] delivery_total_we…
-
### System Info
CPU architecture : x86_64
GPU name : NVIDIA V10 32G
### Who can help?
_No response_
### Information
- [X] The official example scripts
- [ ] My own modified scripts
…
-
AutoGPTQForCausalLM.from_quantized 加载官方4bit量化模型([Llama2-Chinese-13b-Chat-4bit](https://huggingface.co/FlagAlpha/Llama2-Chinese-13b-Chat-4bit/tree/main))报错:NameError: name 'autogptq_cuda_256' is not de…
-
**Describe the bug**
I am trying to fine-tune DeepSeek-Coder-V2-Lite-Instruct (16B) on a system with 8 MI300X GPUs. Running on any number of GPUs less than 8 works as expected and runs to completion. …