Closed Tamannaverma1912 closed 2 years ago
@Tamannaverma1912 you must use the docker --gpus=all flag when launching the Triton container to ensure the docker has access to the GPU. For running on a CPU only system you would want to use the CPU only container. @jbkyang-nvi can you provide the steps for the same?
Hi @Tamannaverma1912 as noted in the instructions here Triton is unable to load any model configuration that requires a GPU.
Additionally, the instructions are a bit outdated since triton-inference-server/docs/examples/model_repository does not work as is. You need to copy an actual CPU only model and put that in the model repository. I recommend doing:
mkdir models
cp -r /Users/tamannaverma/triton-inference-server/docs/examples/model_repository/simple models/
docker run -it --rm -v$PWD/models:/models nvcr.io/nvidia/tritonserver:22.01-py3
tritonserver --model-repository=/models
Hi, I have the same error. How to replicate the scenario.
nvcr.io/nvidia/tritonserver:22.01-py3
on mac M1name: "damage_onnx_batching"
platform: "onnxruntime_onnx"
max_batch_size: 10
dynamic_batching {
# we group batches at least to 100 ms
max_queue_delay_microseconds: 100000
}
input [
{
name: "input_0"
data_type: TYPE_FP32
dims: [3, 640, 640 ]
}
]
output [
{
name: "output_0"
data_type: TYPE_FP32
dims: [9, -1, -1 ]
}
]
model_warmup {
name: "warmup"
batch_size: 2
inputs: {
key: "input_0"
value: {
data_type: TYPE_FP32
dims: 3
dims: 640
dims: 640
random_data: true
}
}
}
instance_group [
{
count: 1
kind: KIND_CPU
}
]
ton-playground-triton-1 | E0302 16:50:00.503077 1 model_repository_manager.cc:1844] Poll failed for model directory 'damage_onnx': unable to get number of CUDA devices: unknown error
triton-playground-triton-1 | I0302 16:50:00.503099 1 model_repository_manager.cc:546] VersionStates() 'damage_onnx'
but I think the problem is on the architecture used when building the image
Hi @fabiofumarola can you share the model you are using? So I can try to repro.
Same issue. Try the command below if you just want to run the demo , which works for me
docker run --rm -p8000:8000 -p8001:8001 -p8002:8002 -v $(pwd)/docs/examples/model_repository:/models nvcr.io/nvidia/tritonserver:22.02-py3 tritonserver --model-control-mode=explicit --load-model simple --model-repository=/models
you can input multiple --load-model
to load more specific models
Hi @jbkyang-nvi I'll share all the repository so that you can take a look. Anyway I've solved by building a docker image using the compose.py on my mac with m1 and specifying the in the docker build command --platform=linux/amd64
Closing issue due to lack of activity. Please re-open the issue if you would like to follow up with this issue.
I have the same issue with m1. Neither I am able to build.
Hi Amarendra,
I’ll send you the link of the built image tomorrow. Best, Fabio
On Mon, 25 Apr 2022 at 06:09, Amarendra Kumar @.***> wrote:
I have the same issue with m1. Neither I am able to build.
— Reply to this email directly, view it on GitHub https://github.com/triton-inference-server/server/issues/3980#issuecomment-1108054518, or unsubscribe https://github.com/notifications/unsubscribe-auth/AAL2SUHSUIIW7LWGSPVBYEDVGYLIBANCNFSM5PHCMW5A . You are receiving this because you were mentioned.Message ID: @.***>
-- Sent from Gmail Mobile
I have the same issue with m1, how to deal with it?
Currently, Triton does not officially support M1 builds.
@fabiofumarola can you share your compose.py
command so future users can try it?
Hi Amarendra, I’ll send you the link of the built image tomorrow. Best, Fabio On Mon, 25 Apr 2022 at 06:09, Amarendra Kumar @.> wrote: I have the same issue with m1. Neither I am able to build. — Reply to this email directly, view it on GitHub <#3980 (comment)>, or unsubscribe https://github.com/notifications/unsubscribe-auth/AAL2SUHSUIIW7LWGSPVBYEDVGYLIBANCNFSM5PHCMW5A . You are receiving this because you were mentioned.Message ID: @.> -- Sent from Gmail Mobile
can you send me the link of the built image?
i would also appreciate a link to the cpu-only M1 build
Yes sorry for the delay. Here you have the release I've built https://hub.docker.com/repository/docker/prometeiads/tritonserver I'll update it with the latest version
Yes sorry for the delay. Here you have the release I've built https://hub.docker.com/repository/docker/prometeiads/tritonserver I'll update it with the latest version
I can't access the repository... I get a 404 :)
do we have a resolution for that? what is the docker image we should use?
Problem Description I was trying to follow the official example starting the server on a cpu-only device by calling the command:
docker run --rm -p8000:8000 -p8001:8001 -p8002:8002 -v/Users/tamannaverma/triton-inference-server/docs/examples/model_repository:/models nvcr.io/nvidia/tritonserver:22.01-py3 tritonserver --model-repository=/models
Here is the logs:
Triton Information Version: 22.01 I am using Mac M1 pro for the local setup.