openlm-research / open_llama

OpenLLaMA, a permissively licensed open source reproduction of Meta AI’s LLaMA 7B trained on the RedPajama dataset
Apache License 2.0
7.29k stars 372 forks source link

Maybe a link on how to set it up ? #6

Closed PGTBoos closed 1 year ago

PGTBoos commented 1 year ago

Might be nice to try this, but how does one set it up in a python jupyter environment ?. (a working step by step would be nice, or an example notebook)

I'm no newbie but never had the time to spend a few days into LLM's. if a 3080ti can run it , ib be fun to give it a try.

young-geng commented 1 year ago

Thanks for the suggestion! Since we train this with JAX using our EasyLM framework, we have detailed documentation for using it with EasyLM. As for using it with PyTorch, unfortunately we weren't able to experiment much as our compute resources are on TPU.

PGTBoos commented 1 year ago

Ok thanks if you would like to draw much attention with it you might reconsider this. (a lot more people would be able to test it, not many people i know of have TPU's these days).


Van: Xinyang (Young) Geng @.> Verzonden: donderdag 4 mei 2023 00:02 Aan: openlm-research/open_llama @.> CC: PGTBoos @.>; Author @.> Onderwerp: Re: [openlm-research/open_llama] Maybe a link on how to set it up ? (Issue #6)

Thanks for the suggestion! Since we train this with JAX using our EasyLM framework, we have detailed documentation for using it with EasyLMhttps://github.com/young-geng/EasyLM/blob/main/docs/llama.md. As for using it with PyTorch, unfortunately we weren't able to experiment much as our compute resources are on TPU.

— Reply to this email directly, view it on GitHubhttps://github.com/openlm-research/open_llama/issues/6#issuecomment-1533809434, or unsubscribehttps://github.com/notifications/unsubscribe-auth/AFQQOG3BIYJ73WMX7S6KJMDXELIVXANCNFSM6AAAAAAXUWZX4M. You are receiving this because you authored the thread.Message ID: @.***>

young-geng commented 1 year ago

You can also use the JAX version on GPU. That's how we test it locally.