-
### Describe the bug
We installed BentoML 0.13.1 as we cannot migrate to 1.x at this moment, the installation is throwing the following error when we are trying to run any bentoml command:
Traceback…
-
## Weekly Goals
- [ ] use ML serving stacks
- [ ] TF serving
- [ ] Pytorch serving
- [ ] bentoML
- [ ] sheldonML
- [ ] KF serving
- [ ] take all helm courses
-
### Describe the bug
If you happen to be using bentoml as a build server involving modular code that you're installing via `pip install -e `, you're going to encounter very strange/bad behavior:
1…
-
-
### Feature request
I exported an onnx model that accept multiple inputs ("input_ids", "input_mask", "input_seg").
And the docs for bentoml(ONNX) only give a simple example, `runner.run.run(tes…
-
### Describe the bug
This keeps happening with mainly transformer based models. However, this error keeps prevailing only after containerizing like using
`bentoml containerize bento_svc:latest` and…
-
### Describe the bug
Hi.
I already created the bentoml model and service and all work as expected.
but cant containerize the bento due to this error:
after run this command: bentoml container…
-
### Describe the bug
benotml serve {my_service}.py:svc --port 3001 is loaded and stdout shows log
```bash
Prometheus metrics for HTTP BentoServer from "{my_service}.py:svc" can be accessed at h…
-
Hi, BentoML team.
This is a new suggestion for Yatai. In general, when serving a ML model, the input provided to the model and the output returned by the model are stored in an external storage, and …
-
### Feature request
Is it possible to rum multimodal LLMs like Qwen VL or LLaVa 1.5 using openllm?
### Motivation
_No response_
### Other
_No response_