NVIDIA-AI-IOT / tf_trt_models

TensorFlow models accelerated with NVIDIA TensorRT
BSD 3-Clause "New" or "Revised" License
684 stars 244 forks source link

Low latency than expected #27

Open paduck86 opened 5 years ago

paduck86 commented 5 years ago

Hello Mr.Jung, Tanks to you, I was able to test faster rcnn with tensorrt.

But, the latency is lower than I expected on my machine. The response time is as follows,

In addition, the memory usage is not much different.

My code is as follows,

trt_graph = trt.create_inference_graph( input_graph_def=frozen_graph, outputs=output_names, max_batch_size=1, max_workspace_size_bytes=1 << 25, precision_mode='INT8' #'FP32' / 'FP16', minimum_segment_size=50 )

Did I do something wrong? I would really appreciate it if you answer for me.

jkjung-avt commented 5 years ago

I put my latest code in my own GitHub repository: https://github.com/jkjung-avt/tf_trt_models. Feel free to check it out.

Meanwhile, I'm not completely sure what your question is. Are you trying to say that TF-TRT fails to optimize 'faster_rcnn_resnet101' at all?