-
```dockerfile
#Base Image
FROM nvcr.io/nvidia/tritonserver:24.04-trtllm-python-py3
USER root
RUN apt update && apt install --no-install-recommends rapidjson-dev python-is-python3 git-lfs curl uuid…
-
**Is your feature request related to a problem? Please describe.**
MONAI supported the TensorRT model conversion since 1.2. However, when tried to export the [diffusion unet](https://github.com/Proje…
-
Hallo,
I have a problem with ssd inception v2 2017-11-17 model.
I want to train using my dataset of 38 classes and port it to UFF and run the inference in TensorRT on Jetson Nano.
I was success…
-
### Search before asking
- [X] I have searched the Ultralytics YOLO [issues](https://github.com/ultralytics/ultralytics/issues) and [discussions](https://github.com/ultralytics/ultralytics/discussi…
-
```
data = tp.iota((3, 3, 2))
data = tp.cast(data, dtype=tp.int64)
print(data)
```
Throws the following error:
```
Traceback (most recent call last):
File "/tripy/debugging_gather.py", l…
-
## Description
I tried to convert the Flux Dit model on L40S with TensorRT10.5, and found that the peak gpu memory exceeded 46068MiB, but 23597MiB gpu memory was occupied during inference. Is this n…
-
### System Info
## Description
I'm am building the DJL-Serving TensorRT-LLM LMI inference container from scratch, and deploying on Sagemaker Endpoints for Zephyr-7B model. Unfortunately, I run i…
-
when using this node, error happens
Error occurred when executing STATIC_TRT_MODEL_CONVERSION:
Exporting the operator 'aten::scaled_dot_product_attention' to ONNX opset version 17 is not support…
-
## ❓ Question
How do you compile a chunk operator with TensorRT? I have been trying a basic example in a Jupyter Notebook but get an unbroadcastable dimension error. The below code executes in PyTo…
-
### System Info
A100 40GB x8, Ubuntu 22.04
### Who can help?
_No response_
### Information
- [ ] The official example scripts
- [ ] My own modified scripts
### Tasks
- [ ] An officially support…