-
### System Info
```shell
+-----------------------------------------------------------------------------+
| HL-SMI Version: hl-1.17.0-fw-51.1.0 |
| Driver Ver…
-
启动命令:
`python api_v2.py -a 0.0.0.0 -p 7860 -c GPT_SoVITS/configs/tts_infer.yaml`
tts_infer.yaml:
`
custom:
bert_base_path: GPT_SoVITS/pretrained_models/chinese-roberta-wwm-ext-large
cnh…
-
Integrate intel/Habana HPU kernel support for uint4 inference on Habana HPU. This was merged into AutoGPTQ https://github.com/AutoGPTQ/AutoGPTQ/pull/689/files but there are no ci tests and we have no …
-
**Feature Overview (aka. Goal Summary)**
Implement Intel Gaudi support in InstructLab project, so Gaudi 2 and Gaudi 3 can be used for SDG, evaluation, and training.
**Goals (aka. expected user out…
ktam3 updated
20 hours ago
-
I want to use model.predict in a loop.
It keeps printing this:
GPU available: True (cuda), used: True
TPU available: False, using: 0 TPU cores
IPU available: False, using: 0 IPUs
HPU available:…
-
### System Info
```shell
HL-SMI Version: hl-1.17.0-fw-51.3.0
Driver Version: 1.17.0-28a11ca
Docker image: vault.habana.ai/gaudi-docker/1.17.0/ubuntu22.04/habanalabs/pytorch-installer-2.3.…
-
### Your current environment
```text
PyTorch version: 2.2.0a0+git8964477
Is debug build: False
CUDA used to build PyTorch: None
ROCM used to build PyTorch: N/A
OS: Ubuntu 22.04.4 LTS (x86_64)
…
-
### Your current environment
```text
Collecting environment information...
WARNING 09-11 21:43:04 _custom_ops.py:14] Failed to import from vllm._C with ModuleNotFoundError("No module named 'vllm.…
-
### Feature request
Current Whisper inference works well with specified language. However, it does not support passing `language=None`, which can detect the language automatically. A `RuntimeError` i…
-
### Anything you want to discuss about vllm.
While implementing disaggregated prefill, we found an error regarding loading weights from safetensors files. We have filed a JIRA ticket[(HS-3164)](https…