-
**Description**
I was unable to build the onnxruntime_backend with OpenVino for Triton Inference Server r22.03 using compatible ONNXRuntime and tensorrt versions (from Triton Inference Server compati…
-
I have created and trained the model whose directory looks something like this -
![image](https://github.com/Azure/azure-sdk-for-python/assets/52820564/c099259a-9cd0-4935-b7f6-4784b531c966)
The ma…
-
inference.memgpt.ai endpoint on error:
File "/usr/local/lib/python3.11/site-packages/requests/models.py", line 1024, in raise_for_status
raise HTTPError(http_error_msg, response=self)
requests.…
-
### 🐛 Describe the bug
Consider a system where a feature service fetches model metadata that has information on what feature to fetch and finally infer from the model. In order for me fetch this me…
-
### Describe the bug
I've gone through all the steps to install Sora and the last step of running gradio/app.py it fails about 2/3 of the way. It hangs on loading shards at 0% and then get the follow…
-
### Operating System
Windows
### Version Information
Recently we have discovered a problem due to the error in the DNN scoring script file. **Please see the workaround in Additional information sec…
-
### 🐛 Describe the bug
While trying to run load tests with latest merged changes on v2 Open inference protocol, I noticed that the example for mnist does not work in preprocessing step. https://git…
-
This would be for SIL Converters
-
**Description**
Currently triton server doesn't capture the full serialized grpc error message in the message field.
proto: https://github.com/triton-inference-server/common/blob/1df32b982a6ed11ead…
-
As described, trying to spin up a `mistralai/Mistral-7B-v0.1` using the examples in the README. This is on an EC2 `g5.xlarge`.
```
import mii
client = mii.serve("mistralai/Mistral-7B-v0.1")
resp…