Closed sungyeon-neubla closed 1 year ago
@sungyeon-neubla Thank you for reporting this issue!
I see you are running 22.07 version of the container. Is it possible to try the newer versions?
Alternatively, is it possible to use --ipc=host
tag for your docker run
command? If yes, could you please try it and see if this helps.
Closing because of the lack of activity. @sungyeon-neubla please share the answers to the above questions if you are still running into the problems.
Description I am experiencing intermittent error with Python BLS Backend Models where it complains that it failed to increase the share memory pool size.
Here is the exact error message from Triton:
Triton Information What version of Triton are you using?
using nvcr.io/nvidia/tritonserver:22.07-py3 docker image
Are you using the Triton container or did you build it yourself?
I am running the Triton Inference Server with Docker where the docker image looks like the following:
Triton Inference Server is brought up using the following command line:
docker run --gpus '"device=2"' --rm --net=host --shm-size=16g -v /<my-home>/triton/models:/models test-triton-server:latest tritonserver --model-repository=/models
To Reproduce Steps to reproduce the behavior. Make inference request to the Python BLS model.
Describe the models (framework, inputs, outputs), ideally include the model configuration file (if using an ensemble include the model configuration file for that as well).
The Python BLS model does the following:
And here is the model configuration:
Expected behavior A clear and concise description of what you expected to happen.
--shm-size=16g should be enough and the error should not occur.