I have trained a model in mmsegmentation. (Pointrend)
I can use this model to inference with jit inference. When I send to inference request to Triton inference server, I got an error.
Model extension is .pt.
I can deploy the model but when I try to send an inference request to the model I got following error:
tritonclient.utils.InferenceServerException: PyTorch execute failure: The following operation failed in the TorchScript interpreter.
Traceback of TorchScript, serialized code (most recent call last):
File "code/__torch__/mmseg/models/segmentors/cascade_encoder_decoder.py", line 93, in forward
grid = torch.unsqueeze(point_coords, 2)
grid0 = torch.sub(torch.mul(grid, CONSTANTS.c1), CONSTANTS.c0, alpha=1)
output1 = torch.grid_sampler(_18, grid0, 0, 0, False)
~~~~~~~~~~~~~~~~~~ <--- HERE
fine_grained_point_feats = torch.squeeze(output1, 3)
grid1 = torch.unsqueeze(point_coords, 2)
pytorch2torchscript.py(122): pytorch2libtorch
pytorch2torchscript.py(187): <module>
RuntimeError: grid_sampler(): expected input and grid to be on same device, but input is on cuda:0 and grid is on cpu
Hello,
I have trained a model in mmsegmentation. (Pointrend)
I can use this model to inference with jit inference. When I send to inference request to Triton inference server, I got an error.
Model extension is .pt.
I can deploy the model but when I try to send an inference request to the model I got following error:
I am also sharing my client script below:
What should add for this error.
Thanks