Closed lss15151161 closed 5 days ago
The multimodal example on backend is on going because the backend of multimodal is very different to pure decoder (like GPU), as mentioned in https://github.com/triton-inference-server/tensorrtllm_backend/issues/463. Close this issue to prevent duplication.
The multimodal example on backend is on going because the backend of multimodal is very different to pure decoder (like GPU), as mentioned in https://github.com/triton-inference-server/tensorrtllm_backend/issues/463. Close this issue to prevent duplication.