shm007g / LLaMA-Cult-and-More

Large Language Models for All, 🦙 Cult and More, Stay in touch !
https://shm007g.github.io/LLaMA-Cult-and-More/
MIT License
427 stars 24 forks source link

Papers #3

Open shm007g opened 1 year ago

shm007g commented 1 year ago

[Instruction tuning with GPT-4, Microsoft, 2023.04]

shm007g commented 1 year ago

[PaLM 2 Technical Report, Google, 2023.05]

shm007g commented 1 year ago

[GPT-4 Technical Report, OpenAI, 2023.03]

shm007g commented 1 year ago

[Sparks of Artificial General Intelligence: Early experiments with GPT-4, MSFT, 2023.04]

shm007g commented 1 year ago

OpenAI Research

InstructGPT: [Training language models to follow instructions with human feedback, OpenAI, 2022.03]

GPT3: [Language Models are Few-Shot Learners, OpenAI, 2020.05]

GPT2

GPT1

other research

https://openai.com/research/techniques-for-training-large-neural-networks https://openai.com/research/sparse-transformer https://openai.com/research/measuring-goodharts-law https://openai.com/research/webgpt https://openai.com/research

shm007g commented 1 year ago

Prompt Tuning

[Prefix-Tuning: Optimizing Continuous Prompts for Generation, 2021/01, Stanford]

[The Power of Scale for Parameter-Efficient Prompt Tuning, 2021/09, Google]

[GPT Understands, Too, 2021/03, Tsinghua, Peking, BAAI]

[P-Tuning v2: Prompt Tuning Can Be Comparable to Fine-tuning Universally Across Scales and Tasks, 2022/03, Tsinghua, BAAI]

shm007g commented 1 year ago

Google Research

T5, Flan-T5

Pathway, UL2, MoE


[LLM are zero-shot rankers for recommender system] [Amazon, textbooks are all you need: learning language representation for sequence recommandation] A new alternative to RLHF just dropped! https://twitter.com/rasbt/status/1663883300522295296 [Direct Preference Optimization: Your Language Model is Secretly a Reward Model https://arxiv.org/abs/2305.18290 ] https://github.com/eric-mitchell/direct-preference-optimization https://github.com/LAION-AI/Open-Assistant/discussions/3347 distilling step by step: outperforming llm with less training data and smaller model size