tensorflow / tensor2tensor

Library of deep learning models and datasets designed to make deep learning more accessible and accelerate ML research.
Apache License 2.0
15.11k stars 3.44k forks source link

Pre-training of Seq2edits for grammar correction task #1896

Open cingtiye opened 2 years ago

cingtiye commented 2 years ago

Description

In this paper, Seq2Edits: Sequence Transduction Using Span-level Edit Operations, I found all tasks are trained for 1M i terations on 170M sentences extracted from English Wikipedia revisions and 176M sentences from English Wikipedia round-trip translated via German. Whether to get tag and position information between source and target sentences during pre-training? Another question about start and end position, target position is used in code rather than souce position. Which position information should be selected?

Thanks!