Closed johnml1135 closed 9 months ago
@isaac091 - can you put in any findings here - why it may or may not work?
Sure. I haven't tried it on just inferencing yet, but I will run those tests.
Results documented in https://github.com/sillsdev/silnlp/pull/308
"Depending on the model and the GPU, torch.compile() yields up to 30% speed-up during inference. To use torch.compile(), simply install any version of torch above 2.0."
Can we do this? What is the drawback?