-
From https://github.com/NeRF-or-Nothing/vidtonerf/issues/74
@PotatoPalooza
Member
[PotatoPalooza](https://github.com/PotatoPalooza) commented [on Mar 1, 2023](https://github.com/NeRF-or-Nothing/v…
-
Hi
Can we use this with Triton inference server model?
-
Hi,
Thank you for your great work. I am trying to build openpose follow your instructions. I tried openpose example in colab, the building process just fail due to inability to connect your server…
-
Description:
The current model ID is not visible in the local API server, making it challenging for users to efficiently create API requests.
Proposed Solution:
Implement a way to view / copy mo…
-
### What would you like to be added?
Currently we have this project https://github.com/sustainable-computing-io/kepler-model-server based in Python that does many things....
Some of that belongs i…
-
Hi,
I've trying to serve different Phi3 models using the Llama.cpp server that is created by the init-llama-cpp ipex.
When I server with this version I have two problems:
1) The server doesn…
hvico updated
2 weeks ago
-
Each time an LLM responds, it also outputs some info about its performance.
```
llama_print_timings: load time = 4732.44 ms
llama_print_timings: sample time = 86.82 ms / 48…
-
If a model fails due to intermittent failures not related to the model itself, it would be nice to have an auto retry.
For example, during the summer we have had some scheduled model failures due to …
-
**Describe the bug**
When model failed to registered because network error, when re-register this model, `sllm-server` reports model is already registered, and cannot remove model by using `sllm-cli …
-
/kind feature
**Describe the solution you'd like**
Currently it is not possible to specify at what path the downloaded model should be available in the model server container. The downloaded model…