Open bhavyajoshi-mahindra opened 6 days ago
我也遇到了这个问题,因为只有一个权重,所以没有model.safetensors.index.json文件。我自己写了一个model.safetensors.index.json文件问题得到解决。
我也遇到了这个问题,因为只有一个权重,所以没有model.safetensors.index.json文件。我自己写了一个model.safetensors.index.json文件问题得到解决。
How exactly did you manage to create your own model.safetensors.index.json file, can you share in detail so that I could create my own file.
How do you load and infer a custom GPTQ quantized Qwen2-VL model (not the default one) using Qwen2VLForConditionalGeneration in WINDOWS
I used the following code.
When I tried, I got the following error :
This are the files that are generated during the quantization
Here is my environment
tokenizers==0.20.3 torch==2.4.1+cu121 torchaudio==2.4.1+cu121 torchvision==0.19.1+cu121 transformers==4.46.2 accelerate==1.1.1 auto_gptq==0.7.1 CUDA == 12.1