Pinaka-ai / SchemaBottleneck

0 stars 0 forks source link

Workable SFT training for SchemaBottleneck #9

Closed nightlessbaron closed 6 months ago

nightlessbaron commented 6 months ago

Description

Add modifications for the supervised fine-tuning aka warm startup for the scheme generator language model. The code is working for T5 (enc-dec architecture). Considerations to make in the future:

  1. If using decoder only LM, make use of causal LMs.
  2. ensure that the complete target is taken into consideration while decoding.

Modifications