Open LeatherDeerAU opened 2 months ago
arch - x86-64 gpu - rtx3070 docker image nvcr.io/nvidia/tritonserver:24.01-trtllm-python-py3 tensorRT-LLM-backend tag - 0.7.2 tensorRT-LLM tag - 0.7.1 (80bc07510ac4ddf13c0d76ad295cdb2b75614618)
@juney-nvidia
examples
models uploaded successfully
sig_fault_logs.txt
Which backend tag should I use for triton-container version 24-01?
related topics (?): https://github.com/triton-inference-server/tensorrtllm_backend/issues/273 https://github.com/NVIDIA/TensorRT-LLM/issues/782 https://github.com/triton-inference-server/tensorrtllm_backend/issues/88
Example models config.pbtxt files: postproccesing.txt preprocessing.txt tensorrt-llm.txt
Could you share all scripts you use?
System Info
arch - x86-64 gpu - rtx3070 docker image nvcr.io/nvidia/tritonserver:24.01-trtllm-python-py3 tensorRT-LLM-backend tag - 0.7.2 tensorRT-LLM tag - 0.7.1 (80bc07510ac4ddf13c0d76ad295cdb2b75614618)
Who can help?
@juney-nvidia
Information
Tasks
examples
folder (such as GLUE/SQuAD, ...)Reproduction
Expected behavior
models uploaded successfully
actual behavior
sig_fault_logs.txt
additional notes
Which backend tag should I use for triton-container version 24-01?
related topics (?): https://github.com/triton-inference-server/tensorrtllm_backend/issues/273 https://github.com/NVIDIA/TensorRT-LLM/issues/782 https://github.com/triton-inference-server/tensorrtllm_backend/issues/88