thu-coai / CommonsenseStoryGen

Implementation for paper "A Knowledge-Enhanced Pretraining Model for Commonsense Story Generation"
102 stars 14 forks source link

Is post-train different from finetune? #6

Open ziweiji opened 3 years ago

ziweiji commented 3 years ago

In your paper, you mentioned that" introduce the knowledge to the pretrained language model by post-training on knowledge-augmented data." In my opinion, post-train is different from finetune. According to paper Post-training for Deep Learning ”propose an extra training step, called post-training, which only optimizes the last layer of the network.“ But in your code, you just finetune GPT2 on the commonsense knowledge.