At first, I'm very happy that this project exists. I could try Beluga2 thanks to the community who shares, like I do, small parts of GPU. That's very impressive!
As a Linux Fedora user, I use Podman instead of Docker. That works exactly the same as Docker in terms of performances.
Hello,
At first, I'm very happy that this project exists. I could try Beluga2 thanks to the community who shares, like I do, small parts of GPU. That's very impressive!
As a Linux Fedora user, I use Podman instead of Docker. That works exactly the same as Docker in terms of performances.
The methods is to follow: https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/latest/cdi-support.html
In short, for Fedora:
Then, launching petals server is easy:
As you can see, the only differences are to set a security option and give the device names.
That works like a charm on my RTX 3070.
Maybe you can add it, or do you need me to create the page / part in the documentation ?