exo-explore / exo

Run your own AI cluster at home with everyday devices 📱💻 🖥️⌚
GNU General Public License v3.0
14.91k stars 803 forks source link

[BOUNTY - $200] Support MLX community models in tinygrad inference engine #200

Open AlexCheema opened 2 months ago

AlexCheema commented 2 months ago
radenmuaz commented 22 hours ago

Does this bounty also requires porting mlx modelling code to tinygrad? Since according to mlx-examples library, different models on mlx-community requires different modelling code. exo currently only has llama. llama tinygrad modelling code is incompatible (different) with weights from qwen, etc.

https://github.com/ml-explore/mlx-examples/blob/bd6d910ca3744d75bf704e6e7039f97f71014bd5/llms/mlx_lm/utils.py#L81

though if models are ported from mlx to tinygrad, we don't need converter anymore.