-
whisper with language set when other language spoken it return translated to the language in option
how to let whisper to not translate and just return blank or ignore the request ?
Thanks
-
Traceback (most recent call last):
File "inference.py", line 97, in
File "utils.py", line 45, in load_checkpoint
File "torch\nn\modules\module.py", line 1672, in load_state_dict
self.__…
-
Problem: We're trying to evaluate C API vs python for online inference. And C API does not have an analogous [thread count](https://github.com/catboost/catboost/blob/master/catboost/python-package/cat…
-
### System Info
I have searched the repo here and the main server repo but don't see any information on either a) support for Safetensors (many models are saved that way on HF) and also b) whether th…
-
When running Gradio cookbook, I am running into this error when trying to execute the very last prompt in the cookbook.
Error message shown in the editor:
`Error Exception: ffmpeg was not found b…
-
**Is your feature request related to a problem? Please describe.**
I would like to use the Intel oneDNN Execution Provider (EP) in ONNX Runtime built for Triton Inference Server ONNX Backend.
**De…
-
**Is your feature request related to a problem? Please describe.**
Currently, when Triton Inference Server is running in `--model-control-mode=explicit` and a `load_model` request is sent from the cl…
teith updated
9 months ago
-
**Describe the bug**
I have folllowed the instructions of CUDA installation. And I have CUDA install on my container also.
But when I tried to run `ilab generate`, it complains.
```
(venv) …
-
**Description**
I use a model ensemble with 3 models: pre-processor, inference model and post-processor. I want to send one image to the server and generate **n** patches of the given image in the pr…
-
i installed pytorch for cpu on debian server
got error:
Pipelines loaded with `torch_dtype=torch.float16` cannot run with `cpu` or `mps` device. It is not recommended to move them to `cpu` or `mps` …