Closed jsjs0827 closed 1 year ago
Can you provide the full log? I have a vague memory that this won't lead to the program crash.
We used TensorRT in Triton server, and got similar error.
http | [W] [TRT] TensorRT was linked against cuDNN 8.4.1 but loaded cuDNN 8.4.0 http | [E] [TRT] 1: [raiiMyelinGraph.h::RAIIMyelinGraph::24] Error Code 1: Myelin (Compiled against cuBLASLt 11.5.2.0 but running against cuBLASLt 11.5.1.0.)
We used TensorRT in Triton server, and got similar error.
I would suggest seeking help from Triton first. this looks like an env issue.
closing since no activity for more than 3 weeks, please reopen if you still have question, thanks!
you got that message because when you convert to TRT, the liblibcublas.so.11.4.2.0
was loaded, but when you do an inference, the lib cuBLASLt 11.5.1.0
was loaded instead oflibcublas.so.11.4.2.0
, you need to install cuda version that has cuBLASLt 11.5.1.0
on your env used for convert model, you could looking at this link to find the cuda version for cuBLASLt 11.5.1.0
cuBLASLt 11.5.1.0 belongs to cuda 11.3.1
Use this code to check what libs loaded during conversion and inference.
def get_loaded_libraries():
import os
import psutil
p = psutil.Process(os.getpid())
for lib in p.memory_maps():
print(lib.path)
how to solve?