The Triton Inference Server provides an optimized cloud and edge inferencing solution.
BSD 3-Clause "New" or "Revised" License
8.18k
stars
1.46k
forks
source link
Python backend "DLPack tensor is not contiguous error" on tensors with a 0 dim #6960
Open
darintay opened 7 months ago
Description pb_utils.Tensor.from_dlpack throws a "DLPack tensor is not contiguous" exception on some (but not all) tensors that have a 0 dimension.
If any dimension is 0, I don't see how the tensor could be non-contiguous.
(Originally I was also having issues with some dim=1 tensors, but it looks like that was fixed in https://github.com/triton-inference-server/python_backend/pull/281)
Triton Information What version of Triton are you using? 23.09
Are you using the Triton container or did you build it yourself? Built self
To Reproduce Here is my simple Python-backend model:
Using torch 1.13.1.
Expected behavior All tensors above should "PASS", as they are all contiguous.