issues
search
nagataka
/
Read-a-Paper
Survey
6
stars
1
forks
source link
RLHF
#50
Open
nagataka
opened
1 year ago
nagataka
commented
1 year ago
Key papers (as of Nov 2023)
Deep reinforcement learning from human preferences
Scalable agent alignment via reward modeling: a research direction
Fine-Tuning Language Models from Human Preferences
Learning to summarize from human feedback
Training language models to follow instructions with human feedback
Llama 2: Open Foundation and Fine-Tuned Chat Models
A General Theoretical Paradigm to Understand Learning from Human Preferences
Key papers (as of Nov 2023)