-
I have a 2 GPU system, a 3060 (12gb VRAM) and a 3070ti (8GB). I've read torch supports paralellism that can split large models into both GPUs, it'd be great to have something like that to run big mode…
-
Can infini run on multi cards? When the requests pressure is high, the bra1 memory is not enough.
![image](https://github.com/user-attachments/assets/7d5f817e-8656-4324-9546-e5c088505c1e)
![image…
-
### Search before asking
- [X] I have searched the Ultralytics YOLO [issues](https://github.com/ultralytics/ultralytics/issues) and [discussions](https://github.com/ultralytics/ultralytics/discussion…
-
Does the device argument support more than one gpu?
rabum updated
4 weeks ago
-
For example from this build: https://buildkite.com/clima/climaatmos-ci/builds/21401#0192def8-fe3a-4986-b3ea-5942e6b9a090
With the error message:
> Executable not found in current directory or st…
-
I have adapted this using simple Data-Parallel from Pytorch, but the model seems to output ``nans sometimes. Have you been able to train this across multiple GPUs on a single node?
-
### Jan version
v0.5.6
### Describe the Bug
When I select just the 4060 GPU, models load onto the 2070.
When I select just the 2070, models load onto the 4060.
### Steps to Reproduce
_No respons…
-
I am currently trying to retrain the BLIP2 architecture on a multi gpu setup using the default torch DDP implementation of the Lavis library.
My training proceeds fine until some steps with consol…
-
### Prerequisites
- [X] I have read the [ServerlessLLM documentation](https://serverlessllm.github.io/).
- [X] I have searched the [Issue Tracker](https://github.com/ServerlessLLM/ServerlessLLM/issue…
-
Currently you can only run one exo instance on each device.
There are some design decisions here:
- Should we support running multiple exo instances on the same device, with one per GPU
- Or shou…