Closed RodriMora closed 2 months ago
You have an older version of the exllamav2 package installed, along with exllamav2_ext
. You can either uninstall the package (pip uninstall exllamav2
) and use the JIT version if you have the CUDA toolkit installed, or you can install the most recent prebuilt wheel from here.
You have an older version of the exllamav2 package installed, along with
exllamav2_ext
. You can either uninstall the package (pip uninstall exllamav2
) and use the JIT version if you have the CUDA toolkit installed, or you can install the most recent prebuilt wheel from here.
you're correct. Thanks!
When trying to quantize llama 3 instruct from HF to exl2 I get this error:
Using this command to quantize:
python convert.py -i /home/ubuntu/text-generation-webui/models/meta-llama_Meta-Llama-3-70B-Instruct -o /home/ubuntu/temp/exl2/ -cf /home/ubuntu/text-generation-webui/models/meta-llama_Meta-Llama-3-70B-Instruct_exl2_8.0bpw -b 8.0
Using the latest version as per git pull of main branch