shm007g / LLaMA-Cult-and-More

Large Language Models for All, 🦙 Cult and More, Stay in touch !
https://shm007g.github.io/LLaMA-Cult-and-More/
MIT License
422 stars 24 forks source link
alpaca chatgpt deepspeed ggml gpt gpt4 gptq llama llm loralib pytorch tensorflow transformers vicuna

Catalog

Pre-trained Base Models

Simple Version - OpenAI: GPT-1, GPT-2, GPT-3, InstructGPT, Code-davinci-002, GPT-3.5, GPT-4(-8k/32k) - Anthropic: Claude-v1, Claude Instant - Meta: OPT, Galactica, LLaMA - huggingface BigScience: BLOOM (176B), BLOOMZ, mT0 - EleutherAI: GPT-Neo, GPT-J (6B), GPT-NeoX (20B), Pythia - TogetherCompute: GPT-JT, RedPajama-7B, RedPajama-INCITE - Berkeley: OpenLLaMA - MosaicML: MPT-7B, MPT-7B-Instruct/Chat - TII: Falcon-7/40B-(instruct) - BlinkDL: RWKV-4-Pile, RWKV-4-PilePlus - Tsinghua THUDM: GLM-130B, ChatGLM-6B - Cerebras: Cerebras-GPT - Google: T5, mT5, LaMDA, Pathways, PaLM, UL2, Flan-T5, Flan-UL2, Bard, PaLM-E, PaLM 2, MoE, Switch Transformer, GLaM, ST-MoE, MoE Routing - DeepMind: Gopher, Chinchilla, Sparrow - Nvidia: Megatron-Turing NLG (530B) - AI21 Studio: Jurassic-1, Jurassic-2 ![A summary of large language models (A Survey of Large Language Models)](https://github.com/hollobit/GenAI_LLM_timeline/assets/998803/9a855dea-7223-4523-924e-3952b1f3734d) ![LLM Family Tree](https://github.com/Mooler0410/LLMsPracticalGuide/blob/main/imgs/qr_version.jpg?raw=true) - [LLMsPracticalGuide](https://github.com/Mooler0410/LLMsPracticalGuide) - [Awesome-LLM](https://github.com/Hannibal046/Awesome-LLM) - [GenAI_LLM_timeline](https://github.com/hollobit/GenAI_LLM_timeline) - [Open-LLMs](https://github.com/eugeneyan/open-llms)

Licences

Open Source Aligned LLMs

$\color{red}{\textsf{Refactoring...}}$

Instruction and Conversational Datasets

Pre-training Datasets

Efficient Training

Efficient Training Library

Typology of Efficient

$\color{red}{\textsf{Refactoring...}}$ raw version here https://github.com/shm007g/LLaMA-Cult-and-More/issues/4

Evaluation Benchmark

Multi-Modal LLMs

$\color{red}{\textsf{Planning}}$

Tool Learning

$\color{red}{\textsf{Planning}}$

Star History