Open marthinwurer opened 13 hours ago
Oh, good catch. Following @gau-nernst's recommendation, I've disabled dynamic compilations, meaning you will hit cache size limits for small caches in the optimizer.\ This can be safely ignored by increasing the cache size limit (128 works well in my runs).
I'll add documentation for that tomorrow.
Increased it to 128 with
import torch._dynamo.config
torch._dynamo.config.cache_size_limit = 128
And now it runs. However, it takes 15-30s for it to compile and run the first batch.
@ClashLuke Btw you can also use torch._dynamo.utils.disable_cache_limit()
within the optimizer. Should be safe I think
Also, did you observe faster speed compared to dynamic-shape compile used previously?
@gau-nernst super cool, wasn't aware of that! I'll add that now, thank you for the pointer
@marthinwurer
However, it takes 15-30s for it to compile and run the first batch.
Hm, yeah, that doesn't seem ideal for prototyping, though the speedups are definitely useful for longer runs.\ I'll expose compile_mode a bit more aggressively
Started my training with a relatively complex resnet autoencoder, got this error.