bigscience-workshop / petals

🌸 Run LLMs at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading
https://petals.dev
MIT License
9.1k stars 510 forks source link

Roadmap (tentative) #12

Open justheuristic opened 2 years ago

justheuristic commented 2 years ago

Current tasks:

End of december: cover more use cases

End of ~july~ august: make it reliable, test with early adopters

End of june: build a proof-of-concept

Important, but not urgent:

justheuristic commented 2 years ago

[moved inference of prompt-tuned model and priorities from summer to current tasks]

bionicles commented 1 year ago

hey, how hard would it be to extend petals to support training these models in addition to the fine tuning?

mryab commented 1 year ago

Hi @bionicles, Petals is a system designed specifically for inference of large models: however, it shares a lot of the underlying architecture with SWARM Parallelism (see https://github.com/yandex-research/swarm for a WIP implementation, which I hope to update in the coming weeks).

The short answer is "definitely possible", but please keep in mind that pretraining is out of scope for Petals. Hence, it might be more useful to continue the discussion elsewhere (e.g. to the SWARM repo or our Discord server) if you have specific questions or suggestions

borzunov commented 1 year ago

Hi @bionicles,

A small addition to the @mryab's response - while Petals does not support training from scratch, both Petals and SWARM are based on hivemind, our library for training over the Internet, which can be used for pre-training. Please see Q3 of the FAQ's "General" section for details.