What steps did you take and what happened:
Basically I saved my yolov8 model as an Onnx model and I tried to follow the documentation for serving.
When I'm trying to do simple inference with internal & external url of the isvc, I get this error
InferenceServerException: [400] Request for unknown model: 'v2' is not found
What did you expect to happen:
Be able to do inference wiht the triton predictor
Checking the logs, I can verify that the onnx model was copied to my mnt/models path and ready on the kserve panel in kubeflow but while looking at the kserve container logs I can't see my model status.
+-------+---------+--------+
| Model | Version | Status |
+-------+---------+--------+
+-------+---------+--------+
I0312 09:18:11.466345 1 tritonserver.cc:1718]
+----------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| Option | Value |
+----------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| server_id | triton |
| server_version | 2.13.0 |
| server_extensions | classification sequence model_repository model_repository(unload_dependents) schedule_policy model_configuration system_shared_memory cuda_shared_memory binary_tensor_data statistics |
| model_repository_path[0] | /mnt/models |
| model_control_mode | MODE_NONE |
| strict_model_config | 1 |
| pinned_memory_pool_byte_size | 268435456 |
| min_supported_compute_capability | 6.0 |
| strict_readiness | 1 |
| exit_timeout | 30 |
+----------------------------------+----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
I0312 09:18:11.470316 1 grpc_server.cc:4111] Started GRPCInferenceService at 0.0.0.0:9000
I0312 09:18:11.470803 1 http_server.cc:2803] Started HTTPService at 0.0.0.0:8080
I0312 09:18:11.513379 1 http_server.cc:162] Started Metrics Service at 0.0.0.0:8002
/kind bug
What steps did you take and what happened: Basically I saved my yolov8 model as an Onnx model and I tried to follow the documentation for serving. When I'm trying to do simple inference with internal & external url of the isvc, I get this error
InferenceServerException: [400] Request for unknown model: 'v2' is not found
What did you expect to happen: Be able to do inference wiht the triton predictor
What's the InferenceService yaml:
Anything else you would like to add:
Checking the logs, I can verify that the onnx model was copied to my
mnt/models
path and ready on the kserve panel in kubeflow but while looking at the kserve container logs I can't see my model status.Environment:
kubectl version
): 1.27/etc/os-release
):