Closed caffeinetoomuch closed 2 years ago
Hi @ice-americano - looks like you are not using a base image that has docker supporting, the image bentoml/model-server:0.13.1-py38
actually does not contain all the cuda and cudnn dependencies that's required. Could you try using the 0.13.1-py38-gpu
image instead and try again?
Hi @ice-americano, can you try this again on release 1.0.0a2
?
Hi @ice-americano, feel free to try out our rc releases with pip install -U --pre bentoml
. I hope this issue should be addressed in recent release of BentoML.
If it is required for you to stay at 0.13, We will come back to this after 1.0 release is out.
BentoML has released official 1.0.0 with Hugging Face Transformers support. Could you please give it a try? Let us know if this problem persists.
I am trying to serve BentoService with GPU by dockerizing it. However, my docker container fails to load model on to GPU. I am using
TransformersModelArtifact
to save and load the model. Docker container runs fine and even handles incoming requests, but model still does not run on GPU. I was able to access GPU inside docker container, though, so it is definitely not a docker issue. There was not any error statements in docker logs.Service definition:
Packing script:
Docker run commands
Environment: