Closed cesumilo closed 4 days ago
For anyone facing the same issue, the problem lies in the model's output. The PyTorch backend cannot handle models that return a dictionary of tensors as output. The solution I found was to use the Python backend to wrap the model and reshape the outputs into tensors. Alternatively, modifying the model's output format could also resolve the issue.
Hi community, I'm sorry if I'm not writing this in the right place. I couldn't figure out where to ask a question about the triton inference server.
Description I have a PyTorch model with the following structure:
The output of the model is a dictionary mapping string keys to lists of tensors.
When I try to run an inference with the following script, I get the error below.
Error:
Triton Information Image:
nvcr.io/nvidia/tritonserver:24.09-py3
Expected behavior I was expecting to get my output tensors from the dictionary by specifying the dictionary structure in the model configuration. I couldn't figure out from the documentation how to make this work.
Any idea how to make this work? 🙏