OpenThaiGPT / openthaigpt-pretraining

Apache License 2.0
21 stars 10 forks source link

[SIIT] Convert MPT to support gradient checkpointing, LoRa #180

Open ArthurMinovsky opened 1 year ago

ArthurMinovsky commented 1 year ago

Description:

Add MPT with Gradient Checkpointing and LoRa support into OpenThaiGPT pertaining code. We will use MPT with Lora for continue pertaining to task #179

To Do:

  1. MPT Weight + MPT with Lora -> Add Lora with PEFT -> Finetune -> Convert to Huggingface checkpoint
  2. Implement MPT Lora with OpenThaiGPT GitHub

Outcomes:

References:

Wintoplay commented 1 year ago

new repository: https://github.com/mikeybellissimo/LoRA-MPT/tree/main

https://colab.research.google.com/drive/1rPwsy659VxOMyx7yUiYo6tsfsg_SyAuf?usp=sharing

still fail: loss == 0

Chawak commented 1 year ago

Resource

Wintoplay commented 1 year ago

https://colab.research.google.com/drive/1qdRMaJ-ro2k7k23d-ICoBq4IjSwiaTnA?usp=sharing

Wintoplay commented 1 year ago

@pitikorn32 (@boss-chanon บอกให้ tag) repo: https://github.com/mikeybellissimo/LoRA-MPT/tree/main

error message after upgrading peft lib: /tmp/tmpfgiig51b/main.c:2:10: fatal error: cuda.h: No such file or directory

include "cuda.h"