mudler / LocalAI

:robot: The free, Open Source alternative to OpenAI, Claude and others. Self-hosted and local-first. Drop-in replacement for OpenAI, running on consumer-grade hardware. No GPU required. Runs gguf, transformers, diffusers and many more models architectures. Features: Generate Text, Audio, Video, Images, Voice Cloning, Distributed, P2P inference
https://localai.io
MIT License
25.11k stars 1.9k forks source link

Global community pools to share LocalAI federated instances and workers #3113

Closed mudler closed 3 months ago

mudler commented 3 months ago

Since now we have federation support #2915 and #2343 it makes sense to build a place under the LocalAI website to list and visualize community pools.

By community pools, I'm refering to a way for people to share swarms token, so they can both provide hardware capabilities, and use the federation for inference (like, Petals, but with more "shards").

The idea is to have an "explorer" or a "dashboard" which shows a list of active pools and with how many federated instances or llama.cpp workers, reporting their capability and availability.

Things to notice:

This would allow the users to:

1) setup a cluster, and dedicate that for a specific community 2) share the compute resources with others 3) run inferencing if you don't have beefy hardware, but it is instead given by other community peers

mudler commented 3 months ago

This would likely be a new golang app that can be deployed e.g. in Vercel and would need a simple form for users to submit tokens.

I see two sections in this app:

1) a page or a form to insert new tokens and provide description/name 2) a landing page where to show all the global pools, with the availability, number of workers, and hardware specs ( note this is not yet collected by the p2p swarm functionalities)

mudler commented 3 months ago

thinking it again. no need of vercel or a dynamic web app at all: can all be static and have GH workflow pipelines to run "cron" jobs to update the data.

mudler commented 3 months ago

thinking it again. no need of vercel or a dynamic web app at all: can all be static and have GH workflow pipelines to run "cron" jobs to update the data.

scratch that - too complicated to add new tokens then

mudler commented 2 months ago

https://explorer.localai.io is now live. It still misses some of UX around on how to run things, but it's just low hanging fruit on documentation that will be addressed in follow-ups