Open fmichaelobrien opened 1 year ago
see - https://github.com/ObrienlabsDev/blog/wiki/CUDA-based-%E2%80%90-High-Performance-Computing-%E2%80%90-LLM-Training-%E2%80%90-Ground-to-GCP-Cloud-Hybrid#use-cases Start with the following site - read it in its entirety https://jaykmody.com/blog/gpt-from-scratch/
or https://github.com/lm-sys/FastChat
review https://github.com/vectara/hallucination-leaderboard
Thinking also of RTX-A6000 with 48G (nvlink) - as an alternative to my two RTX-A4500s with 20+20 = 40G under an nvlink
see - https://github.com/ObrienlabsDev/blog/wiki/CUDA-based-%E2%80%90-High-Performance-Computing-%E2%80%90-LLM-Training-%E2%80%90-Ground-to-GCP-Cloud-Hybrid#use-cases Start with the following site - read it in its entirety https://jaykmody.com/blog/gpt-from-scratch/
or https://github.com/lm-sys/FastChat