OpenThaiGPT / openthaigpt-pretraining

Apache License 2.0
21 stars 10 forks source link

Continue Pretraining Llama7B with Huggingface trainer #310

Closed boss-chanon closed 11 months ago

boss-chanon commented 11 months ago

Information in requirement: https://docs.google.com/document/d/1DQ5a56gFv2ZzRKn9PliNlOc-d-CI5vGAUUsdPihX4OI/edit#heading=h.5bb1u3i0p4rh

Additional Steps: 1 Understand Lanta Multinode training guidelines first https://app.gitbook.com/o/ygzlt6vZbi4mM0I2X5ko/s/rXqu9ENRkozaiYy0LTZK/lanta/multinode-training

  1. Try to integrate multinode training with the Huggingface Trainer script
  2. Integrate our own Huggingface dataset (V5_555) into Training code
  3. Run all Checklist
  4. Notify @kwan or @boat or @new before running full training
  5. Run Training Note: This step is blocked by Task:Integrate Data Pipeline into OSCAR Colossal

Important Note: No need for fancy config script, or codebase. Just one train.py (200-400 lines) is enough for this task. Recommend modifying Standford Alpaca training code from Multinode training guidelines

Recommended People to do the task (Pick 1)

  1. @New
  2. @Boss
  3. @Boat
  4. @Bank
  5. @Tae