-
/kind bug
**What steps did you take and what happened:**
I ran the inference service on custom xgboost model that I trained and saved in .joblib extension using the pvc storage option, followed th…
-
This PR is part of an effort to improve integration of feast with model serving. Also see #4139 and accompanying draft [RFC](https://docs.google.com/document/d/1PzBbTs_8R73XhuDq3CO0slmGy5S_ci2rwtbx1L-…
-
### Search before asking
- [X] I have searched the YOLOv8 [issues](https://github.com/ultralytics/ultralytics/issues) and [discussions](https://github.com/ultralytics/ultralytics/discussions) and fou…
-
### System Info
Image: v1.2 CPU
Model used: jinaai/jina-embeddings-v2-base-de
Deployment: Docker / RH OpenShift
### Information
- [X] Docker
- [ ] The CLI directly
### Tasks
- [X] An officiall…
-
### Problem
For now I have to create some gateway via AWS or other supported cloud to start any service with dstack.
But if I have on-prem pool of hardware, I actually don't need to publish my infer…
-
### Feature request
Currently the Service Name for OTLP is hard-coded as "text-generation-inference.server"
Could an environment variable be added which could set this. Something like...
resour…
-
### Application contact emails
sachidesai@microsoft.com, guofei@microsoft.com, ishaansehgal@microsoft.com, jpalma@microsoft.com, qike@microsoft.com
### Project Summary
KAITO automates the dep…
-
Since the interface seems unnecessary for batch automatic labeling, could you provide a way to operate without a GUI?
-
/kind bug
**What steps did you take and what happened:**
Deployed inferenceservice iris-classifier-deployment:
```
% kubectl get inferenceservices
NAME URL …
-
**Description**
After using the Python vllm backend, Triton crashed with signal 11. The model had been loaded and preheated for some time before the crash occurred.
**Triton Information**
What ve…