AkihikoWatanabe / paper_notes

たまに追加される論文メモ
https://AkihikoWatanabe.github.io/paper_notes
17 stars 0 forks source link

Few-Shot Data-to-Text Generation via Unified Representation and Multi-Source Learning, ACL'23 #863

Open AkihikoWatanabe opened 1 year ago

AkihikoWatanabe commented 1 year ago

https://virtual2023.aclweb.org/paper_P4950.html

AkihikoWatanabe commented 1 year ago

In this paper, we present a novel approach for data-to-text generation that addresses the limitations of current methods that primarily focus on specific types of structured data. Our proposed method aims to improve performance in multi-task training, zero-shot and few-shot scenarios by providing a unified representation that can handle various forms of structured data such as tables, knowledge graph triples, and meaning representations. We demonstrate that our proposed approach can effectively adapt to new structured forms, and can improve performance in comparison to current methods. For example, our method resulted in a 66\% improvement in zero-shot BLEU scores when transferring models trained on table inputs to a knowledge graph dataset. Our proposed method is an important step towards a more general data-to-text generation framework.

Translation (by gpt-3.5-turbo)