Closed HtutLynn closed 4 years ago
Do you see the same problem if you start with a clean copy of the latest tensorrt_demos code?
Have you tested other models, say, yolov3-tiny-288 or yolov4-288? Are you seeing the same problem too?
Hi, sorry for the late reply.
Hi,
Nvidia GTX 1070
Try updating NVIDIA driver on this machine.
Has this issue been resolved?
@jkjung-avt , I am sorry for the late reply. I haven't updated the Nvidia driver on my machine because I am kinda afraid of breaking things so I tested on Jetson Nano 4GB and it works fine as expected. I get around the same performance as you reported in YOLOv4 performance table section when I test with yolov4-288.trt . However, when I try to do inference with yolov4-tiny-288.trt on Nano, the performance is only around ~9FPS compared to the performance on the table, 36.6FPS. That's a pretty big difference. Jetson Nano is flashed with Jetpack 4.4 and trt_yolo.py is run on a container from NGC. I followed every step that's mentioned on README so I am not sure what is causing this performance difference.
If you set up your Jetson Nano the same way as how I did it, you should get similar results as mine. I documented how I set up my Jetson Nano DevKit with JetPack-4.4 in the JetPack-4.4 for Jetson Nano blog post. In particular, the versions/configurations of OpenCV and "protobuf" might affect the result. And there could be certainly other stuffs which cause differences.
Hi, The same problem happened in my setting, "Tesla K40m, TensorRT-7.0.0.11, Driver Version: 418.39 CUDA Version: 10.1". Thanks for your reply
The same problem happened in my setting, "Tesla K40m, TensorRT-7.0.0.11, Driver Version: 418.39 CUDA Version: 10.1". Thanks for your reply
Referring to TensorRT Release 7.x.x, "Kepler" (CUDA Compute 3.x) has been deprecated in TensorRT 7.x. Please test with a more modern GPU.
The same problem happened in my setting, "Tesla K40m, TensorRT-7.0.0.11, Driver Version: 418.39 CUDA Version: 10.1". Thanks for your reply
Referring to TensorRT Release 7.x.x, "Kepler" (CUDA Compute 3.x) has been deprecated in TensorRT 7.x. Please test with a more modern GPU.
Sorry,I have no more modern GPU, maybe I should try TensorRT 6.x or TensorRT 5.x. By the way, I checked the caffe samples in "TensorRT-7.0.0.11/samples/python/introductory_parser_samples", the results seems to be ok.
By the way, I checked the caffe samples in "TensorRT-7.0.0.11/samples/python/introductory_parser_samples", the results seems to be ok.
I think yolov4 model uses some layers which are not present in the "introductory_parser_samples" (resnet50). If you don't have a newer GPU for testing, I think you should verify the result with TensorRT 6 instead.
@jkjung-avt , Anyone solved? I felt this project is the most friendly tutorials of trt use, but I found that I have tested on Ubuntu18.04 x86-64 gtx1060, trt 7.1.3.4 + cuda11.0 + cudnn8. almost the latest GPU environments, but it still no result shown QAQ.
I would still suggest to update your NVIDIA driver. I use the following versions on my RTX-2080 Ti and GTX-1080 and it works.
@jkjung-avt Amazing! it works, thanks a lot~
Hi, I've been working with your old codebase for YOLOv4 TensorRT model conversion and stuff. I just pulled the codebase for the latest commits and tried to convert YOLOv4 tiny 288 to TensorRT with exact steps described and do inference with it but the model is not detecting anything. After that, I checked outputs of the TensorRT runtime model and the inferred NumPy array turns out to be just zeros.
Is there any fix to it? I've been trying to deploy the YOLOV4-tiny-288 TensorRT model on Jetson Nano 2GB.