issues
search
triton-inference-server
/
client
Triton Python, C++ and Java client libraries, and GRPC-generated client examples for go, java and scala.
BSD 3-Clause "New" or "Revised" License
517
stars
224
forks
source link
issues
Newest
Newest
Most commented
Recently updated
Oldest
Least commented
Least recently updated
Fix typo
#687
matthewkotila
closed
1 month ago
1
Fix HTTP Client Async Code for Request Rate
#686
dyastremsky
closed
1 month ago
1
fix: Get validation outputs by name rather than index
#685
rmccorm4
closed
3 weeks ago
1
fix: Fix Client Http Async Code for Request Rate
#684
nnshah1
closed
3 weeks ago
0
Arbitrary Json extra input
#683
IzzyPutterman
closed
3 weeks ago
1
Incomplete installation of all genai-perf dependencies prevents its from being run on air-gapped servers
#682
mirekphd
opened
1 month ago
2
Refactor the Statistics data source
#681
debermudez
closed
3 weeks ago
1
Improve async request rate for OpenAI HTTP client
#680
dyastremsky
closed
1 month ago
7
Add dummy client
#679
Tabrizian
opened
1 month ago
0
Any example of triton-vllm in c++?
#678
tricky61
closed
1 month ago
0
Unit tests for artifacts produced by GenAI-Perf
#677
lkomali
closed
3 weeks ago
3
Document how to profile multiple LoRA adapters
#676
dyastremsky
closed
1 month ago
0
feature: triton generate support
#675
nnshah1
opened
1 month ago
0
Fast PA teardown
#674
tgerdesnv
opened
1 month ago
0
Fast PA teardown
#673
tgerdesnv
closed
1 month ago
0
Set add special tokens to false
#672
IzzyPutterman
closed
1 month ago
7
Patch vLLM for missing content entry
#671
IzzyPutterman
closed
1 month ago
0
Revert "Speed up GenAi-Perf's help call"
#670
dyastremsky
closed
1 month ago
0
Speed up GenAi-Perf's help call
#669
dyastremsky
closed
1 month ago
1
fix is_contiguous_logic issue in dlpack
#668
wjddn279
opened
1 month ago
1
vLLM ITL fix
#667
IzzyPutterman
closed
1 month ago
3
Update the module dependency (#663)
#666
tanmayv25
closed
1 month ago
0
Ci updates
#665
nvda-mesharma
opened
1 month ago
0
Add compare subcommand to README
#664
nv-hwoo
closed
1 month ago
0
Update the module dependency
#663
tanmayv25
closed
1 month ago
0
MultiLoRA Support
#662
IzzyPutterman
closed
1 month ago
0
input_data
#661
riyajatar37003
closed
1 month ago
1
add json exporting
#660
debermudez
closed
1 month ago
0
Add support for FetchContent or find_package
#659
philipp-schmidt
opened
1 month ago
1
Fix request-rate request order
#658
tgerdesnv
closed
1 month ago
0
urllib dependency is present when using [grpc, cuda] options
#657
bgedik
opened
1 month ago
0
Add a check if ModelParser pointer is nullptr (#652)
#656
debermudez
closed
1 month ago
0
Initialize logger for unit tests
#655
dyastremsky
closed
1 month ago
0
Upate GenAI version
#654
mc-nv
closed
1 month ago
0
Fix version for GenAI-Perf
#653
mc-nv
closed
1 month ago
0
Add a check if ModelParser pointer is nullptr
#652
lkomali
closed
1 month ago
0
Fix test_inference_profiler test
#651
lkomali
closed
1 month ago
0
Draft: WIP: Tokens to TRTLLM backend
#650
IzzyPutterman
opened
1 month ago
1
Update r24.05.
#649
mc-nv
closed
1 month ago
0
Revert urllib3 version pin
#648
tokoko
opened
1 month ago
10
R24.05 cherry pick
#647
ganeshku1
closed
1 month ago
0
Fix response preprocessing bug
#646
nv-hwoo
closed
1 month ago
1
Filter model name with slash for artifact path
#645
nv-hwoo
closed
1 month ago
3
Converting InferenceRequest to InferInput
#644
ZhanqiuHu
opened
1 month ago
0
Support Python 3.8 in GenAI-Perf
#643
dyastremsky
closed
1 month ago
5
Changes to support Ensemble Top Level Response Caching (#560)
#642
lkomali
closed
1 month ago
0
Update backend name in tutorial
#641
dyastremsky
closed
1 month ago
2
Add automation to run unit tests and check code coverage for GenAI-Perf
#640
dyastremsky
closed
1 month ago
0
Update scaling from ns to ms
#639
debermudez
opened
1 month ago
2
Add compare subcommand for visualizing multiple profile runs
#638
nv-hwoo
closed
1 month ago
0
Previous
Next