Build your own GPT2 quickly, without doing many useless work.
This project is base on 🤗 transformer. This tutorial show you how to train your own language(such as Chinese or Japanese) GPT2 model in a few code with Tensorflow 2.
You can try this project in colab right now.
├── configs
│  ├── test.py
│  └── train.py
├── build_tokenizer.py
├── predata.py
├── predict.py
└── train.py
git clone git@github.com:mymusise/gpt2-quickly.git
cd gpt2-quickly
python3 -m venv venv
source venv/bin/activate
pip install -r requirements.txt
this is a example of raw dataset: raw.txt
python cut_words.py
python build_tokenizer.py
python predata.py --n_processes=2
python train.py
python predict.py
ENV=FINETUNE python finetune.py