Open zou3519 opened 2 years ago
As it pertains to 1.13:
import torch
time (on my machine) goes from 1.1s with importing the decomps to 1.0s withoutNot sure if that's significant (10%?) but an easy fix is to lazily load the decomps for jvp the first time forward-mode AD is called
That sounds pretty reasonable and seems easy to do indeed
https://github.com/pytorch/pytorch/pull/85989 resolved the immediate item for 1.13, though we should still reconsider the design because it uses TorchScript
We landed https://github.com/pytorch/pytorch/pull/84976. What it does is:
Unfortunately this has been causing some problems:
import torch
. Distributed folks may be sensitive to that.Alternatively, it would not be too difficult to directly call the Python (from C++) instead of relying on TorchScript to shepherd the code through. This may be a better design in the long-term, the question is, should we do anything about this issue in the short term.
cc @soulitzer @malfet