Open bltcn opened 3 years ago
Hi! I haven't tested image on windows. Have you checked container logs?
Preparing models...
[04:56:39] INFO - Preparing 'glintr100' model...
[04:56:39] INFO - Building TRT engine for glintr100...
[TensorRT] WARNING: onnx2trt_utils.cpp:362: Your ONNX model has been generated with INT64 weights, while TensorRT does not natively support INT64. Attempting to cast down to INT32.
[TensorRT] WARNING: Detected invalid timing cache, setup a local cache instead
[TensorRT] WARNING: GPU error during getBestTactic: Conv_0 : invalid argument
[TensorRT] ERROR: 10: [optimizer.cpp::computeCosts::1855] Error Code 10: Internal Error (Could not find any implementation for node Conv_0.)
Traceback (most recent call last):
File "prepare_models.py", line 54, in
Have you tried running other GPU based containers on wsl2, like TensorFlow benchmarks, to verify your wsl2 is properly configured for GPU usage?
Try running this sample: https://docs.nvidia.com/cuda/wsl-user-guide/index.html#ch05-sub01-simple-containers
Run "nbody -benchmark [-numbodies=
NOTE: The CUDA Samples are not meant for performance measurements. Results may vary when GPU Boost is enabled.
Windowed mode Simulation data stored in video memory Single precision floating point simulation 1 Devices used for simulation GPU Device 0: "Pascal" with compute capability 6.1
Compute 6.1 CUDA device: [NVIDIA GeForce GTX 1060] 10240 bodies, total time for 10 iterations: 8.868 ms = 118.245 billion interactions per second = 2364.896 single-precision GFLOP/s at 20 flops per interaction
Hm, then TensorRT should work as expected.
I can double check that latest published version of InsightFace-REST works out of the box, but unfortunately I can't help you with running it on Windows.
I have checked building from scratch with clean clone from repo - everything works as intended on Ubuntu 20.04.
Looks like it's WSL related problem.
thanks, I have tested cpu version. it works fine.maybe there is somthing wrong with parameters in this case
Quote from Nvidia page above:
With the NVIDIA Container Toolkit for Docker 19.03, only --gpus all is supported.
This might be the case, since deploy_trt.sh
tries to set specific GPU. Try replacing line 99 with --gpus all
Though according to the same document there also might be issues with pinned memory required for TensorRT, and issues with concurrent CUDA streams.
If pinned memory is also the issue you can try add RUN $PIP_INSTALL onnxruntime-gpu
to Dockerfile_trt
and switch inference backend to onnx
in deploy_trt.sh
at line 105
thanks. i will try
Hi! Any updates? Have you managed to run it under WSL2?
sorry,I just see your reply. I will try.
sorry,I just see your reply. I will try.
Looks like WSL2 just wasn't supported by TensorRT, but according to change log latest TensorRT version should support it. Try using 21.12 TensorRT image.
sorry,I just see your reply. I will try.
Looks like WSL2 just wasn't supported by TensorRT, but according to change log latest TensorRT version should support it. Try using 21.12 TensorRT image.
i try 21.12 and 22.01TensorRT image, unfortunately,all failed. 21.12 report GPU error during getBestTactic, 22.01 report Cuda failure: integrity checks failed
i try 21.12 and 22.01TensorRT image, unfortunately,all failed. 21.12 report GPU error during getBestTactic, 22.01 report Cuda failure: integrity checks failed
Have you tried running other GPU based containers on WSL2 to ensure everything is installed correctly?
win11,wsl2,ubuntu18.04 How can I deal with it?