turboderp / exllama

A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.
MIT License
2.74k stars 215 forks source link

llama_cpp_python_cuda is not a supported wheel on this platform #300

Closed arif599 closed 1 year ago

arif599 commented 1 year ago

Hey guys,

I'm getting this error when trying to install exllama dependencies, I'm getting this error

ERROR: llama_cpp_python_cuda-0.2.7+cu117-cp310-cp310-manylinux_2_31_x86_64.whl is not a supported wheel on this platform.

Any idea how I can fix it?

Using CentOS Linux 7 (Core)

arif599 commented 1 year ago

Fixed using https://github.com/oobabooga/text-generation-webui#amd-metal-intel-arc-and-cpus-without-avx2