Open Actis92 opened 2 years ago
Recently PyTorch team integrated Torch-TensorRT into the Pytorch ecosystem. blog post
Any tips on how would one implement an export_trt
to Trainer?
We could follow the pattern used by to_onnx
: https://github.com/Lightning-AI/lightning/blob/0ca3b5aa1b16667cc2d006c3833f4953b5706e72/src/pytorch_lightning/core/module.py#L1798. Comparing it to the snippet in your linked blogpost, the advantage would be to automatically use self.example_input_array
(if defined) and call the batch transfer hooks to apply any transformations (if defined). This is what the top post also suggests.
@rohitgr7 Hi is there any progress on this? I want to do super fast GPU inference with my model trained in PyTorch Lightning. How do we convert it to TRT and will it speedup inference 2x or 4x? Thanks, Sam
@davodogster would you be interested to take it over and implementing it? :rabbit:
Hi @Borda ! Sorry, I am an applied data scientist and not a good developer so it may be a challenge for me.
Do you thinks it's easily possible for me to convert my lightning model (image segmentation, batch size >=8) to RensorRT for 3-5x speedup for inference?
š
š Feature
Add a method like to_torchscript in lightgning.py that allow to convert a model in TorchTensorRT in order to increase performance
Motivation
Increase performance during inference
Proposal
Additional context
A possible problem could be the dependencies because it depends on CUDA, cuDNN and TensorRT as you can see https://nvidia.github.io/Torch-TensorRT/v1.0.0/tutorials/installation.html and some of these dependencies I think work only on Linux
cc @borda @carmocca @awaelchli @ninginthecloud @daniellepintz @rohitgr7