Open brent-halen opened 5 months ago
Not 100% sure what you've tried, but perhaps your docker image only has CUDA runtime installed and not CUDA development files? You could try adding a build step using one of Nvidia's "devel" docker images where you compile llama-cpp-python and then copy it over to the docker image where you want to use it.
Or if it's a good enough solution, the easiest method is probably to use a pre-compiled wheel with CUDA support, e.g. https://github.com/jllllll/llama-cpp-python-cuBLAS-wheels
I'm attempting to install llama-cpp-python under the tensorflow-gpu docker image (nightly build) . When I attempt to do so, I get the following error messages.
I attempted to make sure the CUDA bin files were in the path & LD_LIBRARY_PATH variables, but this doesn't seem to have mitigated the issue.
Here's the docker command I use to create the image.
Any advice as to what I should do differently would be greatly appreciated.