-
As a model developer, I want to register my model in reppo protocol so that I can track usage, as well as earn tokens for modeling and revenue from usage.
Acceptance criteria:
- A Reppo ID is assoc…
-
Hi, I cannot submit inference results of ref-youtube-vos to the server:
https://competitions.codalab.org/competitions/29139#participate-submit_results
Is this server out of service? And does this me…
-
Hi,
My config: A770 + Ollama + OpenWebui + intelanalytics/ipex-llm-inference-cpp-xpu:latest docker
After 2-3 chat message I get this error:
```ollama_llama_server: /home/runner/_work/llm.cpp/llm.…
-
### Issues Policy acknowledgement
- [X] I have read and agree to submit bug reports in accordance with the [issues policy](https://www.github.com/mlflow/mlflow/blob/master/ISSUE_POLICY.md)
### Where…
-
Currently using multiple GPU to eval, the model will load into serveral devices, caused inference error.
How to force them eval one task in parallel on multiple device?
-
Hi this is JeongSoo Na
After installation in README file, I try this command
But this error is printed
```
(dynamicbind) PS C:\Users\jsna\Desktop\Github\DynamicBind> python .\run_single_pr…
-
I don't plan on shelling out money for inference at the moment, so the initial plan is to have users bring their own "inference back-end" with them - likely Colab for now. Some points about this thoug…
-
**Description**
`MetricModelReporter` is not populating available GPU information to appropriate metric labels.
**Triton Information**
Latest `main`, commit `28fe1d119686a4378bf081a1b7cca068631ee…
-
I ran the infernce of Falcon-7b and neural-chat-7b-v3-1 models on ray server with below command
python inference/serve.py --config_file inference/models/neural-chat-7b-v3-1.yaml --simple
python infe…
-
**Description**
I am trying to analyse input_buffer (Json input) to aim how actually the "TYPE_STRING" inference request is getting converted into byte arrays by Triton .
**Triton Information**
W…