issues
search
triton-inference-server
/
server
The Triton Inference Server provides an optimized cloud and edge inferencing solution.
https://docs.nvidia.com/deeplearning/triton-inference-server/user-guide/docs/index.html
BSD 3-Clause "New" or "Revised" License
8.37k
stars
1.49k
forks
source link
issues
Newest
Newest
Most commented
Recently updated
Oldest
Least commented
Least recently updated
fix: Fixing pip installation as a system package
#7768
KrishnanPrash
closed
2 weeks ago
0
docs: Add clarification for label_filename in classification docs
#7766
trevoryao
opened
2 weeks ago
0
PyTorch model with Dictionary[Key,Tensor] output
#7765
cesumilo
closed
4 days ago
1
Fix model generation
#7764
mc-nv
closed
2 weeks ago
0
Build: Fixing model generation
#7763
pvijayakrish
opened
2 weeks ago
0
Do I need to warm up the model again after reloading it?
#7762
soulseen
opened
2 weeks ago
1
How to deploy ensemble models of different versions more elegantly?
#7761
lzcchl
opened
2 weeks ago
1
StatusCode.UNAVAILABLE] Received http2 header with status: 502
#7760
furkanc
opened
2 weeks ago
0
build: Install tritonfrontend and tritonserver wheels by default in published containers
#7759
KrishnanPrash
closed
2 weeks ago
1
Build Updates for Ubuntu24.04
#7758
nv-anants
closed
2 weeks ago
0
Reapply "Change compute capability min value (#7708)"
#7757
mc-nv
closed
2 weeks ago
0
ci: Adding tests for `numpy>=2`
#7756
KrishnanPrash
closed
2 weeks ago
0
Unable to simultaneously load TensorRT model.plan on different GPUs in Triton Inference Server in the same instance
#7755
AntnvSergey
opened
3 weeks ago
0
Build: Update main branch post 24.10 release
#7754
pvijayakrish
closed
3 weeks ago
0
docs: Simplify PR templates
#7753
yinggeh
opened
3 weeks ago
0
test: Test model_metrics config and document histogram buckets override
#7752
yinggeh
closed
2 weeks ago
1
refactor: Include job id and nightly tag to results uploaded
#7751
kthui
closed
3 weeks ago
0
Release: Update README for r24.10
#7750
pvijayakrish
closed
3 weeks ago
0
ensemble logic control
#7749
xiazi-yu
closed
1 week ago
1
This topic describes python_backend memory usage
#7748
1120475708
closed
3 weeks ago
0
CI: Update test script for TRT compatibility test
#7747
pvijayakrish
closed
3 weeks ago
0
wget: unable to resolve host address ‘contentmamluswest001.blob.core.windows.net’
#7746
ST-q
closed
3 weeks ago
1
Build AMD64 Triton from ARM64 machine generate ARM64 architecture executable file
#7745
ti1uan
opened
3 weeks ago
2
build: RHEL8 Python Backend
#7744
fpetrini15
closed
3 weeks ago
0
feat: Enable deferred unregistering of shared memory regions after inference
#7743
pskiran1
opened
4 weeks ago
0
Handle raw binary request in python
#7741
remiruzn
opened
4 weeks ago
0
SeamlessM4T on triton
#7740
Interwebart
opened
4 weeks ago
0
Expensive & Volatile Triton Server latency
#7739
jadhosn
opened
4 weeks ago
1
About automatic Batch
#7738
CallmeZhangChenchen
closed
3 weeks ago
1
Running multi-gpu and replicating models
#7737
JoJoLev
opened
4 weeks ago
1
fix: Copy models out of NFS before starting Triton to avoid intermitt…
#7736
indrajit96
closed
4 weeks ago
0
test: Add delay to L0_lifecycle test_load_new_model_version after each model file update
#7735
kthui
closed
4 weeks ago
2
Fix array size for jetson tests
#7734
mc-nv
closed
4 weeks ago
0
Custom Image build doesn't detect Debian system
#7733
VishDev12
opened
1 month ago
1
Error building Triton Docker image in CPU-Only mode with TensorFlow2 backend
#7732
PierreCarceller
opened
1 month ago
1
Failing CPU Build
#7731
coder-2014
closed
1 day ago
2
fix: Copy models out of NFS before starting Triton to avoid intermittent startup timeouts
#7730
rmccorm4
closed
1 month ago
0
chore: ensure proper clean up in shared memory related tests
#7729
GuanLuo
closed
3 weeks ago
0
Triton is stopping, unexpectedly and without logging, when using a large model, s3 and periodical checks to ready and live endpoints
#7728
smcbn
opened
1 month ago
0
Memory Leak in NVIDIA Triton Server (v24.09-py3) with model-control-mode=explicit
#7727
Mustafiz48
opened
1 month ago
6
Unrecognized configuration class to build an AutoTokenizer for microsoft/Florence-2-base-ft
#7726
shihao28
closed
3 weeks ago
1
Build: Update Openvino and vLLM versionsfor Release 24.10
#7725
pvijayakrish
closed
1 month ago
0
No content returned with OpenAI-Compatible Frontend Beta
#7724
Loc8888
closed
4 weeks ago
1
Caught signal 11 (Segmentation fault: address not mapped to object at address 0x1c0)
#7723
wxk-cmd
opened
1 month ago
1
Facing import error in python backend on Apple M2/M3 chips
#7722
TheMightyRaider
opened
1 month ago
3
Revert "Change compute capablity min value (#7708)"
#7721
mc-nv
closed
1 month ago
0
refactor: moving `tritonfrontend` to `@handle_triton_error` decorator
#7720
KrishnanPrash
closed
1 month ago
0
ONNX CUDA session not working in python backend
#7719
jsoto-gladia
opened
1 month ago
3
[Bug] Error when serving Torch-TensorRT JIT model to Nvidia-Triton
#7718
zmy1116
opened
1 month ago
0
Removing caching on windows.
#7717
mc-nv
closed
3 weeks ago
2
Previous
Next