-
How can I run inference and get .png segmentation mask on pascal voc test set, which is neede for pascal voc server submission?
-
**Is your feature request related to a problem? Please describe.**
I'd like to be able to run vLLM emulating the OpenAI compatible API to use vLLM as a drop-in replacement of ChatGPT.
**Describe…
-
hello,
I want to just inference of pre-trained model in the terminal, but I don't want to run a HTTP server. How could I do that?
-
### What would you like to be added?
Currently we have this project https://github.com/sustainable-computing-io/kepler-model-server based in Python that does many things....
Some of that belongs i…
-
### The bug
I set "quota_immich" in the OAuth setting and also as user attribute in Keycloak, but in immich the quota doesnt set right on first login.
On debug mode in the log is the claim passthr…
-
I forked your wonderful program and updated the requirements.txt and a few other files to now work with python 3.10 and Pytorch 2.2.1.
I also fixed some of the warnings for DataLoader and the Pandas…
-
### What happened?
Version on all 3 Machines:3978 (ff252ea4) I just used Git pull and updated to all 3 machines and then rebuilt the cmake.All the machines are on ethernet.I am not sure if it is n…
-
I have a bert model that I am trying to deploy with Triton Inference Server using Tensorrt-LLM backend. But I am getting errors:
? Docker Image: 24.03
? TensorRT-LLM: v0.8.0
Error:
+-------+-…
-
### System Info
## System Specifications
2024-11-10T21:20:44.880890Z INFO text_generation_launcher: Runtime environment:
Target: x86_64-unknown-linux-gnu
Cargo version: 1.80.1
Commit sha: https:/…
-
**Description**
PR [185](https://github.com/triton-inference-server/client/pull/185) pinned `geventhttpclient==2.0.2` due to a potential change in ssl_context_factory handling.
The geventhttpcli…