:robot: The free, Open Source alternative to OpenAI, Claude and others. Self-hosted and local-first. Drop-in replacement for OpenAI, running on consumer-grade hardware. No GPU required. Runs gguf, transformers, diffusers and many more models architectures. Features: Generate Text, Audio, Video, Images, Voice Cloning, Distributed inference
There is a problem if you try to connect aio-cpu images to aio-gpu images and vice versa. It's looked like we can only have CPU networks and GPU networks, because if you launch local-ai-gpu as host and aio-cpu as worker, then the host will try to assemble backend on worker side and catch CUDA errors, because worker device have no gpu.
The same thing possibly work backwards, meaning if you have aio-cpu as a host and aio-gpu as a worker then GPU instance will get backend with cpu-only output.
LocalAI version:
local-aio-gpu-nvidia-cuda-12 local-aio-cpu
Environment, CPU architecture, OS, and Version:
Describe the bug
I am trying to build p2p network, and it's actually working, peers can discover each other and exchange tasks (https://localai.io/features/distribute/)
There is a problem if you try to connect aio-cpu images to aio-gpu images and vice versa. It's looked like we can only have CPU networks and GPU networks, because if you launch local-ai-gpu as host and aio-cpu as worker, then the host will try to assemble backend on worker side and catch CUDA errors, because worker device have no gpu.
The same thing possibly work backwards, meaning if you have aio-cpu as a host and aio-gpu as a worker then GPU instance will get backend with cpu-only output.
To Reproduce
Start local-ai-gpu-nvidia-cuda-12 as host node and local-ai-cpu as worker node
Expected behavior
Host side would understand that it's cpu instance and will not try to build gpu backend
Logs
Additional context