Open iKora128 opened 1 month ago
I have seen the same issue with Python 3.12.4 on Windows, using CUDA 12.5.
Any resolution for this? I have cuda 12.1, python 3.10, and onnxruntime-gpu 1.18.1. I can see the libonnxruntime_providers_cuda.so is still looking for libcublasLt.so.11, was this by design? Shouldn't this be working with cuda 12?
Running into this while using runpod, but works fine on Colab with the 1.17.1 workaround. @iKora128 any chance you were hitting this on Runpod as well?
What happened?
Problem It seems that while onnxruntime-gpu is running with CUDA 12, when I try to use fastembed, it references CUDA 11, resulting in an error and not working.
[Environment]
[Content] I followed the installation page for onnxruntime-gpu (https://onnxruntime.ai/docs/install/) and downloaded the wheel as well. While onnxruntime-gpu works fine, when I attempt to use fastembed for embedding, I encounter the following issue:
[onnxruntime-gpu Operation Check Code]
Results:
[Error Code]
Error:
It seems that fastembed is referencing CUDA 11 for some reason. Since onnxruntime-gpu appears to recognize the GPU without issues, I'm not sure what the problem is.
If there is any mistake in my setup, I apologize. Does anyone else have the same issue or know if there is a mistake in the setup process? Any help would be greatly appreciated.
What Python version are you on? e.g. python --version
Python 3.11.6
Version
0.2.7 (Latest)
What os are you seeing the problem on?
Linux
Relevant stack traces and/or logs
No response