Closed HZapperz closed 1 month ago
ilab serve
uses llama-cpp-python for inference. You have to build and install llama-cpp-python with CUDA support.
@HZapperz perhaps this will help?
pip cache remove llama_cpp_python
pip install --force-reinstall llama_cpp_python==0.2.75 -C cmake.args="-DLLAMA_CUBLAS=on"
pip install instructlab -C cmake.args="-DLLAMA_CUBLAS=on"
https://developer.ibm.com/tutorials/awb-installing-instructlab-on-a-gaming-pc Also mentioned here https://developer.ibm.com/tutorials/awb-synth-train-contribute-instructlab-submission/
This issue has been automatically marked as stale because it has not had activity within 90 days. It will be automatically closed if no further activity occurs within 30 days.
This issue has been automatically closed due to inactivity. Please feel free to reopen if you feel it is still relevant!
Discussed in https://github.com/instructlab/instructlab/discussions/1312