Closed PGTBoos closed 1 year ago
Thanks for the suggestion! Since we train this with JAX using our EasyLM framework, we have detailed documentation for using it with EasyLM. As for using it with PyTorch, unfortunately we weren't able to experiment much as our compute resources are on TPU.
Ok thanks if you would like to draw much attention with it you might reconsider this. (a lot more people would be able to test it, not many people i know of have TPU's these days).
Van: Xinyang (Young) Geng @.> Verzonden: donderdag 4 mei 2023 00:02 Aan: openlm-research/open_llama @.> CC: PGTBoos @.>; Author @.> Onderwerp: Re: [openlm-research/open_llama] Maybe a link on how to set it up ? (Issue #6)
Thanks for the suggestion! Since we train this with JAX using our EasyLM framework, we have detailed documentation for using it with EasyLMhttps://github.com/young-geng/EasyLM/blob/main/docs/llama.md. As for using it with PyTorch, unfortunately we weren't able to experiment much as our compute resources are on TPU.
— Reply to this email directly, view it on GitHubhttps://github.com/openlm-research/open_llama/issues/6#issuecomment-1533809434, or unsubscribehttps://github.com/notifications/unsubscribe-auth/AFQQOG3BIYJ73WMX7S6KJMDXELIVXANCNFSM6AAAAAAXUWZX4M. You are receiving this because you authored the thread.Message ID: @.***>
You can also use the JAX version on GPU. That's how we test it locally.
Might be nice to try this, but how does one set it up in a python jupyter environment ?. (a working step by step would be nice, or an example notebook)
I'm no newbie but never had the time to spend a few days into LLM's. if a 3080ti can run it , ib be fun to give it a try.