Open 0-chan-kor opened 11 months ago
Hi - thank you for the question. Currently, there is not a way to export/serialize an artifact from torch.compile
. Our ir="dynamo"
path does have serialization capabilities, however, and could be helpful here (it can be invoked via torch_tensorrt.compile(..., ir="dynamo",...)
. In torch.compile
, the generated TRT Engines are stored for the duration of the Python sessions and should not need recompilation for additional inference calls within the same session, but between Python sessions, we do not yet have a caching/saving mechanism.
❓ Question
Hello. I am currently trying to infer the stable diffusion XL inpaint model using your package. model link : https://huggingface.co/diffusers/stable-diffusion-xl-1.0-inpainting-0.1
I referred to your example code and modified it as follows.
On my gpu machine the conversion to tensorrt takes over 15 minutes. Since I can't do this conversion every time, I'm trying to find a way to save it in file format such as ".trt" file and use it.
When looking in your documentation, it was difficult to find such a feature. Do you support these features? If so, please let me know.
What you have already tried
Described above
Environment
docker container : nvcr.io/nvidia/pytorch:23.11-py3 gpu : p40
Additional context