-
### The bug
After upgrading from v1.115, my external library assets show up in trash as offline. However, they can not be restored and still show as offline after a library re-scan. The server was up…
-
Hi,
I'm using MLServer with KServe, and found that the proto descriptor in grpc has a collision between them:
```
File ~/.cache/pypoetry/virtualenvs/example-mlflow-lZ2hGP5g-py3.10/lib/python3.10/…
-
问题类型:模型部署
训练过程中保存的模型可以推理,但导出模型后载入出错。
**导出模型**
```
(paddle2.1) liuyu@ai-Super-Server:~/jli/paddlexs$ paddlex --export_inference --model_dir=./output/faster_rcnn_r50_fpn/best_model --save_dir=./…
-
Add reference code for `mixtral-8x7b`(https://github.com/mlcommons/inference/tree/master/language/mixtral-8x7b) in `axs`.
To do the following steps:
- add recipe for downloading dataset
- add recip…
-
python models_server.py --config configs/config.default.yaml # required when `inference_mode` is `local` or `hybrid`
python awesome_chat.py --config configs/config.default.yaml --mode server # for te…
-
I've converted a model to tensorRT format. And at the inference time, I need to initialize the state of the model with a custom value. I implemented this with pycuda in python inference. but it was no…
-
**Is your feature request related to a problem? Please describe.**
Currently I can't define the log4j properties because the default log4j properties is hardcoded in.
https://github.com/aws/sage…
-
## Description
When using official mxnet docker image on cloud server and without docker on a local server, the inference result is different. The input is exactly the same.
## Environment info (R…
-
Similar to the work performed [langchain-llm-api](https://github.com/1b5d/langchain-llm-api) I would like to see the ability to use this natively within langchain. Are there any plans to do so such th…
-
#### Description
I am currently working on deploying the Seamless M4T model for text-to-text translation on a Triton server. I have successfully exported the `text.encoder` to ONNX and traced it …