-
Hi Marcos,
You mentioned a future update with classification model support in your Readme and I was wondering if we could have some information about this update (estimated release date, difficult…
-
### Feature request
Does any documentation exist, or would it be possible to add documentation, on how to use the TensorRT-LLM backend? #2458 makes mention that the TRT-LLM backend exists, and I can …
-
There's no way to stop an inference. Reloading or closing the webpage doesn't stop it either. The nodes just keep going.
-
### System Info
- Architecture: x86_64
- OS Ubuntu 22.04
- GPU: NVIDIA GeForce RTX 4090
- Gpu memory 2x24gb
- CPU max MHz: 5000.0000
- Driver Version: 535.183.01
- CUDA Version: 12.2
- Conta…
-
**Description**
All gRPC requests to the Triton server are timing out, but HTTP requests are functioning normally.
**Triton Information**
23.10
Are you using the Triton container or did you bu…
-
Hi,
I have an extensive background in dotnet stack however for about 1 year and half I have been using python to perform all things machine learning. For last few months I have been studying bayesi…
-
can't generate SDXL refiner model engines.
PS: SDXL base+refiner workflow,How should `sd_unet` be configured?
Exporting sd_xl_refiner_1.0_0.9vae to TensorRT
{'sample': [(2, 4, 128, 128), (2, 4, 1…
-
-
It would be great if there are examples of using Android or iOS devices.
-
I am working with the Tensorflow 2.0 project that uses multiple models for inference.
Some of those models were optimized using TF-TRT.
I tried both regular offline conversion and offline conversi…