The Triton Inference Server provides an optimized cloud and edge inferencing solution.
BSD 3-Clause "New" or "Revised" License
8.38k
stars
1.49k
forks
source link
feat: Metrics Support in `tritonfrontend` #7703
Closed
KrishnanPrash closed 3 weeks ago
What does the PR do?
Adding support for
Metrics
intritonfrontend
. This involves two components:HTTPMetricsServer
Metrics
classWith this PR, similar to
KServeHttp
andKServeGrpc
, the metrics service can used with:Additional Changes made in this PR:
request.post(...)
based on this commentTest plan:
Added 3 test function to L0_python_api:
test_metrics_default_port()
: Tests whether the metrics service can start as expectedtest_metrics_custom_port()
: Tests whether arguments defined intritonfrontend.Metrics.Options
are passed successfully toHTTPMetrics
test_metrics_update()
: Tests whethernv_inference_count
value goes from 0 to 1 if inference request is performed.