awslabs / multi-model-server

Multi Model Server is a tool for serving neural net models for inference
Apache License 2.0
998 stars 230 forks source link

Issue: Memory Leak when serving multiple models #999

Open pratikluitel opened 2 years ago

pratikluitel commented 2 years ago

Description

I am encountering a memory leak when serving multiple MXNet models behind the same endpoint in multi-model-server.

I am using 2 docker containers, with the multi-model-server docker image, and serving 4 models in each container. Here are the relevant parts of my docker compose file: image

The issue

There is massive memory leak. One would expect the memory to clear after each inference, but it keeps on adding and adding until the multi-model-server stops.

This issue does not occur when I use separate containers to serve each model, serving one model per container, like so: image

Only 500MB memory is consumed per model in this case, which does not increase at all on multiple inferences. But when serving multiple models, each inference uses extra memory, and the memory does not clear at all. The multi model server crashes after it runs out of memory.

chinge55 commented 2 years ago

I have encountered this as well. +1

chinge55 commented 2 years ago

In my case, even when I separated the models into multiple containers; the memory was still leaking.

pratikluitel commented 2 years ago

update: seems like this is an issue with multi model server itself. there was no memory leak when serving these models using a flask server. Hope this is fixed soon

lrbsunday commented 1 year ago

The issue seems still there. Is this tool still in maintenance?