Open matbee-eth opened 3 weeks ago
Sorry, the current code doesn't support multi-gpus. We will optimize these functions in the future. For inference time, you can refer to https://github.com/VectorSpaceLab/OmniGen/blob/main/docs/inference.md#requiremented-resources
Oh bummer, I have 2x 1070, and just realized it only using one GPU. Keep up the good work and hope to see more optimization and can use low VRAM GPU mode.
Sorry, the current code doesn't support multi-gpus. We will optimize these functions in the future. For inference time, you can refer to https://github.com/VectorSpaceLab/OmniGen/blob/main/docs/inference.md#requiremented-resources
Hey, @staoxiao will this thing make it work?
https://stackoverflow.com/questions/54216920/how-to-use-multiple-gpus-in-pytorch
I have 2x 4090, is it possible to have this model run across multiple gpus, or offload the kv cache to another cuda device? I'm fairly certain its the kv_cache being offloaded to cpu thats causing it to be so slow.