-
- Serving multiple ML models using mlflow in a single VM [stackoverflow](https://stackoverflow.com/questions/70620074/serving-multiple-ml-models-using-mlflow-in-a-single-vm)
- How to get stable resul…
-
ARG VALUE
__________________________________________________
ckpt_name = bert_model.ckpt
config_name = bert_config.json
cors = *
…
-
### Configuration
I'm trying to update a model using databricks_model_serving. The first time it works, but when I try to deploy another version the error `Error: cannot update model serving: scaleTo…
-
**Prerequisites**
> Please fill in by replacing `[ ]` with `[x]`.
* [x] Are you running the latest `bert-as-service`?
* [x] Did you follow [the installation](https://github.com/hanxiao/bert-as-…
-
### Your current environment
```text
Collecting environment information...
/data/miniconda3_new/envs/vllm-new/lib/python3.10/site-packages/transformers/utils/hub.py:127: FutureWarning: Using `TRANS…
-
### Your current environment
```text
```
### How would you like to use vllm
I have three 4090 GPUs with a total of 24*3 GB of memory, and the model I need to deploy requires at least 52 GB. Th…
-
Hello,
Firstly, thank you for this repo. When I try to connect the Ollama node to the Mistral7B model which is locally served by Ollama serve. I am getting this error again and again. Is using the …
-
**What did you find confusing? Please describe.**
I tried to submit an invoke endpoint request to a serverless endpoint deployed with sagemaker that uses one of these containers. however I can only f…
-
**Describe the bug**
downloadOllama.js windows absolute url 404's.
**To Reproduce**
+ First time installing - installed as admin
+ Open Reor and immediately get this error message:
```
Error: …
-
### Feature request
We want to understand how serving models might be useful in the context of OpenAdapt.
e.g. serve llama.cpp via LMTP with https://github.com/eth-sri/lmql/tree/main/src/lmql/mo…