Closed ShuaiShao93 closed 6 months ago
That would be great, especially given the torch cache
Thanks for your feature request. I've linked your issue to this ticket and will close this issue as duplicate. https://github.com/triton-inference-server/server/issues/3547
Let's keep all the comments related to this feature in that GitHub issue.
Is your feature request related to a problem? Please describe. I know we can copy GPU tensor to CPU with
torch.utils.dlpack.from_dlpack
(link), but we don't want to introduce torch into our deps.Describe the solution you'd like Add an API to Tensor to copy to CPU and convert to numpy.