nktice / AMD-AI

AMD (Radeon GPU) ROCm based setup for popular AI tools on Ubuntu 22.04 / 23.04 / 23.10 / 24.04
Other
134 stars 8 forks source link

Llama-cpp-python compilation for text-generation-webui using LLAMA_HIPBLAS=on for rocm #7

Open mpwx007 opened 1 week ago

mpwx007 commented 1 week ago

For those who have some problems compiling llama-cpp-python with ROCm 6.0 (https://github.com/abetlen/llama-cpp-python), here is how I did it:

NB: If your version is different, edit it accordingly.

Hope this help you !

nktice commented 1 week ago

I went and tried this... and found another method making it from source.

conda activate textgen
# remove old version if needed...
pip uninstall llama_cpp_python
pip uninstall llama_cpp_python_cuda

mkdir ~/text-generation-webui/repositories/
cd ~/text-generation-webui/repositories/
git clone  --recurse-submodules  https://github.com/abetlen/llama-cpp-python.git
cd llama-cpp-python
pip install .