-
As is mentioned in this [issue](https://github.com/NVIDIA/TensorRT-LLM/issues/110) that the release branch does not support the bfloat16+weight_only_int8 quantization, while this feature is already su…
-
Hello, I want to optimize model by convert model to TensorRT
But I'm not familiar with pytorch, so I don't know how to do it.
Do I need to modify the entire code to apply tensorrt optimization t…
-
@CarkusL Thanks for your great work.I Merge pfe_sim.onnx and rpn.onnx to pointpillars_trt.onnx. And use it by TensorRt to inference.But the result is error which is showed in the link. could you help…
-
I need to deploy my model at some point. Does anyone have an answer? Thanks!!
-
### 🐛 Describe the bug
When I compile with USE_TENSORRT=1 I get the following error:
onnx2trt_utils.hpp:30:10: fatal error: onnx/onnxifi.h: No such file or directory
Note that removing USE_TENSOR…
-
请问一下 paddlehub 里的 bert 模型能支持 TensorRT 6 的预测吗~
另外能找到的教程都是 C++ 的 ,请问 Paddle-TensorRT 库有实现 python 的接口吗
-
I realized that the Tensorflow Lite does not support inference with using Nvidia GPU. I have a device of Nvidia Jetson Xavier. My current inference is made with unoptimized transformers model on GPU. …
-
Hello,jaybdub.
Thanks for you code.
I am trying to accelerate the trained Pytorch model inference process with Tensorrt
but got nothing useful from searched blogs.
I found you have done this work …
-
Hello @CarkusL I have tried you method for converting centerpoint into onnx and then into tensorRT. Is there a way can visualize the engine file output or calculating the validation accuracy of the en…
-
goog job!