Open jlewi opened 1 month ago
I was able to work around this by doing the following
Create a symbolic link
ln -s /usr/local/lib/python3.10/dist-packages/tensorrt_llm/libs/libnvinfer_plugin_tensorrt_llm.so /usr/lib/libnvinfer_plugin_tensorrt_llm.so.9
Set the LD_LIBRARY_PATH as follows
export LD_LIBRARY_PATH=/usr/local/lib/python3.10/dist-packages/tensorrt_llm/libs:/usr/local/nvidia/lib64:/opt/tritonserver/backends/tensorrtllm:/opt/tritonserver/lib
If I didn't set the LD_LIBRARY_PATH I got errors about not being able to find a bunch of different libraries
I'm running on GKE. I believe libcuda.so.1
is provided by the driver and gets installed on the host. That might explain why it gets installed in a location. that the triton server image doesn't know about and requires explicit configuration. I'm not sure about the others.
this seems to be solved by adding
RUN ldconfig
System Info
CPU Architecture x86 A100 40GB
Who can help?
No response
Information
Tasks
examples
folder (such as GLUE/SQuAD, ...)Reproduction
Expected behavior
I expected the server to start.
actual behavior
I get an error
additional notes
The library
libnvinfer_plugin_tensorrt_llm.so
is availableI have set