Closed MrWaggel closed 3 months ago
Unsure why memory increases, but you could instead call predictor.persist()
and keep the predictor in memory, that would also speed up the inference calls significantly.
Please open a new issue referencing this one if the issue persists in the latest version of AutoGluon (v1.1+)
Describe the bug After running
TabularPredictor.predict(...)
successively serially not parallel, memory isn't being freed in a consistent manner. On production server with limited memory capacity it hits OOM quite fast.In my case after I run prediction on one row of 248 bytes of data.
note Not that familiar with python, so I might be doing something completely wrong. In any case if someone could point me in the right direction on how to solve this.
Expected behavior Return allocated memory for garbage collection. No dangling references in other threads for garbage collection?
Installed Versions
Additional context Selected model was trained with preset
high_quality
(bagged/stacked).