Open mpwx007 opened 1 week ago
I went and tried this... and found another method making it from source.
conda activate textgen
# remove old version if needed...
pip uninstall llama_cpp_python
pip uninstall llama_cpp_python_cuda
mkdir ~/text-generation-webui/repositories/
cd ~/text-generation-webui/repositories/
git clone --recurse-submodules https://github.com/abetlen/llama-cpp-python.git
cd llama-cpp-python
pip install .
For those who have some problems compiling llama-cpp-python with ROCm 6.0 (https://github.com/abetlen/llama-cpp-python), here is how I did it:
_(/!\ your HIP_VISIBLEDEVICES can be 1, check it if it not working) [check if there is no error with the following command:
/opt/rocm/bin/hipconfig --full
] 2:sudo apt-get install libstdc++-12-dev libstdc++-12-doc
3:conda activate textgen
4:CMAKE_ARGS="-D LLAMA_HIPBLAS=ON -D CMAKE_C_COMPILER=/opt/rocm-6.0.2/llvm/bin/clang -D CMAKE_CXX_COMPILER=/opt/rocm-6.0.2/llvm/bin/clang++ -D CMAKE_PREFIX_PATH=/opt/rocm-6.0.2 -D AMDGPU_TARGETS=gfx1100" pip install llama-cpp-python==0.2.75 --no-deps --force-reinstall --no-cache
NB: dont forget to change gfx1100 if you have other graphic card than RX 7900 serieNB: If your version is different, edit it accordingly.
Hope this help you !