-
Can infini run on multi cards? When the requests pressure is high, the bra1 memory is not enough.
![image](https://github.com/user-attachments/assets/7d5f817e-8656-4324-9546-e5c088505c1e)
![image…
-
推理时,使用文档中的语句:
```bash
python inference.py --asr hubert --dataset ./your_data_dir/ --audio_feat your_test_audio_hu.npy --save_path xxx.mp4 --checkpoint your_trained_ckpt.pth
```
![image](https://…
-
Right now GPU implements Vulkan, Metal, D3D12 and D3D11 backends. Out of these APIs D3D11 is the odd one out because it has awkward support for command buffers. Why do we support it?
The main reaso…
-
This issue has emerged multiple times on discord
https://discourse.julialang.org/t/memory-usage-increasing-with-each-epoch/121798
https://discourse.julialang.org/t/flux-memory-usage-high-in-srcnn/…
-
[The new Skia](https://github.com/aseprite/aseprite/issues/139) port supports GPU acceleration (Direct3D, OpenGL, Metal). Anyway the GPU acceleration isn't working as expected (it's slower than the so…
dacap updated
7 hours ago
-
![image](https://github.com/user-attachments/assets/267e35b9-b6a2-411c-80d3-6b90492ad6db)
As per https://cloud.google.com/free/docs/free-cloud-features
Your Free Trial credits apply to all Google …
-
### Your current environment
The output of `python collect_env.py`
```text
device=V100 * 4
vllm==0.6.4.post1
```
### 🐛 Describe the bug
When I attempt to use the newly added…
-
I am trying to use both of my GPUs who are passed through to my docker container.
```
services: faster-whisper-server-cuda: image: fedirz/faster-whisper-server:latest-cuda build: dockerfile: Dockerf…
-
### Describe the bug
When I run the test case “tests/integrate/102_pw_pint_uks” using the command OMP_NUM_THREADS=1 mpirun -n 1 abacus, I noticed that if I set both `init_wfc nao` and `device gpu`, t…
-
My use case scenario is deploying model inference services in the cloud, utilizing GPU virtualization technology to split one GPU into multiple instances. Each instance runs a model, and since one car…