Open justheuristic opened 2 years ago
[moved inference of prompt-tuned model and priorities from summer to current tasks]
hey, how hard would it be to extend petals to support training these models in addition to the fine tuning?
Hi @bionicles, Petals is a system designed specifically for inference of large models: however, it shares a lot of the underlying architecture with SWARM Parallelism (see https://github.com/yandex-research/swarm for a WIP implementation, which I hope to update in the coming weeks).
The short answer is "definitely possible", but please keep in mind that pretraining is out of scope for Petals. Hence, it might be more useful to continue the discussion elsewhere (e.g. to the SWARM repo or our Discord server) if you have specific questions or suggestions
Hi @bionicles,
A small addition to the @mryab's response - while Petals does not support training from scratch, both Petals and SWARM are based on hivemind, our library for training over the Internet, which can be used for pre-training. Please see Q3 of the FAQ's "General" section for details.
Current tasks:
End of december: cover more use cases
End of ~july~ august: make it reliable, test with early adopters
End of june: build a proof-of-concept
Important, but not urgent: