-
the instructions make no mention as to HOW to install the gpu drivers when creating the VM. So I just put them in the folder, but I didn't add them to the copyfilestoVM file and didn't add them to any…
-
### Checklist
- [X] The problem is not listed in the [hardware support matrix](https://github.com/ilya-zlobintsev/LACT?tab=readme-ov-file#hardware-support) as a known limitation. Examples of known dr…
-
### 🐛 Describe the bug
Currently, building Intel GPU for Windows requires sourcing Intel GPU development bundle for Windows to setup development environment for PyTorch. But the bundle contains `li…
-
### Your current environment
```text
Collecting environment information...
WARNING 07-23 19:11:42 _custom_ops.py:14] Failed to import from vllm._C with ModuleNotFoundError("No module named 'vllm.…
-
Hi Kyle Champley,
Thanks a lot for this amazing toolkit. It is very convenient to use with Pytorch. However, I am facing an issue that my GPU memory always increases when I iteratively use FBP on d…
-
Thanks for this nice repo.
**Describe the bug**
Finetuning in 2:4 sparsity w4a16 example fails with multiple GPUs
**Expected behavior**
The finetuning step expected to train successfully with …
-
Hello,@jishengpeng . I'm testing your job, but I found the code for wavtokenizer.decode can't be inferred in GPU. Could you tell me how I can fix this problem?
-
### Description
On `trunk-minor`:
- [x] Deprecate `gpu_ids`.
- [x] Replace with `gpu_id`.
On `trunk-major`:
- [x] Remove the multi-GPU code.
### Motivation and context
Multi-GPU was specifica…
-
We want to remove the both CPU and GPU launch latency from GPU kernel runtime. Right now `do_bench` uses CUDA event, which means it will include the GPU launch latency.
It would be more accurate to…
-
It would be helpful to have an option to specify which GPU to use when running inference on a machine with multiple GPUs. In my case, I am running multiple MONAILabel servers, each with its own dedica…
che85 updated
1 month ago