-
I still can not understand which option( w_elem_format_bp, a_elem_format_bp, a_elem_format_bp_ex, a_elem_format_bp_os ) represents gradient?
In fact , in the BP process, I wish to set the gradient as…
-
### System Info
GPU name (NVIDIA A6000)
TensorRT-LLM tage (v0.9.0 main)
transformers tage (0.41.0)
### Who can help?
@nc
### Information
- [X] The official example scripts
- [X] My own modified…
-
Can you add this model for generative fill?
https://github.com/black-forest-labs/flux
-
Hi - while I was waiting for you to fix the Pulid FLux - I thought I would try the Maker, dual setting with SDXL base - but I got this error
![image](https://github.com/user-attachments/assets/640552…
-
### Expected Behavior
I expect no issues. I had installed comfyui anew a couple days ago, no issues, 4.6 seconds per iteration~
### Actual Behavior
After updating, I'm now experiencing 20 seconds p…
-
I'm using an RTX 4090 GPU to run a 5B model, but I keep getting out-of-memory errors. I'm using the cogvideox_5b_example_01 workflow from the example. What could be the reason?
![image](https://githu…
-
### System Info
CPU architecture: x86_64
Host RAM: 1TB
GPU: 2xL20 SXM
Container: Manually built container with TRT 9.3 Dockerfile.trt_llm_backend
TensorRT-LLM version: 0.12.0.dev2024070200"
Dr…
-
### Your question
it was working fine yesterday but now I am having this error... I don't know why, it's my first time using an image-generation model so I don't know what to do. it is working fine…
-
### What behavior of the library made you think about the improvement?
I have just started to use Outlines, and my use case is that I am hosting a local model on a server using [Serve with vLLM](http…
-
Trying to replicate the benchmark by following [the official guide](https://nvidia.github.io/TensorRT-LLM/performance.html) for Llama2-7b with latest release `0.7.1` and triton server image `23.12-trt…