-
When I am running 'construct_data.py', I find the code is running at cpu which leads to very slow processing. However can I run this code at GPU?
I have carefully check the availability of the gpu …
-
Random error
observed once while running on colab T4 instance
-
Sorry for dumb question. When I follow the instructions [here](https://github.com/facebookresearch/faiss/blob/main/INSTALL.md#step-1-invoking-cmake) with `-DFAISS_ENABLE_C_API=ON`, I obtain `libfaiss_…
-
### Is there an existing issue for this problem?
- [X] I have searched the existing issues
### Operating system
Windows
### GPU vendor
Nvidia (CUDA)
### GPU model
4090
### GPU VRAM
24
### Ve…
-
### System Info
- `transformers` version: 4.46.2
- Platform: Linux-5.15.0-1049-oracle-x86_64-with-glibc2.35
- Python version: 3.10.14
- Huggingface_hub version: 0.25.1
- Safetensors version: 0.…
-
### Your current environment
Below is my current Docker Compose configuration:
```yaml
services:
vllm:
image: vllm/vllm-openai:v0.6.4
deploy:
resources:
reservation…
-
0ry0n updated
2 months ago
-
I have 2 4090 and I want to merge 8 7B models. But I get out of memory.
And only one GPU is used. So, how to use 2 4090 simultaneously.
Or there is other method to solve this?
-
### Problem Description
Loading a shared object containing gpu code that is not compiled for one of the gpus in the system causes CLR to assert here https://github.com/ROCm/clr/blob/65d174c3e35423bf2…
-
**Reported by davea on 7 Aug 44506374 20:47 UTC**
Allow GPU exclusive apps to turn off computation on a subset of GPUs
Migrated-From: http://boinc.berkeley.edu/trac/ticket/1361