EleutherAI / project-menu

See the issue board for the current status of active and prospective projects!
65 stars 4 forks source link

[Project] Preparatory-Training (Placeholder name) #32

Closed zphang closed 1 year ago

zphang commented 3 years ago

tl;dr: Add a second phase of pre-training to explicitly train models to be good at prompt-tuning and few-shot learning.

Full proposal: https://docs.google.com/document/d/1ER4tCfXO_Qn6GBWLi70sjrz-6kaMj8HCrTmFFnLf088/

Short version:

  1. Choose a pretrained model e.g. GPT-J, T5, GLM
  2. Build a large set of NLU and NLG tasks. Piggy-back off https://github.com/EleutherAI/lm_evaluation_harness/ where possible.
  3. Do a second phase of pretraining (important: has to be pretraining scale, not fine-tuning scale) on massive task dataset, incorporating prompt tuning during training
  4. Evaluate!
StellaAthena commented 2 years ago

@zphang To what extent would you say that this was scooped by FLAN and T0?