tensorflow / tensor2tensor

Library of deep learning models and datasets designed to make deep learning more accessible and accelerate ML research.
Apache License 2.0
15.4k stars 3.48k forks source link

wikisum uses transformer decoder or full transformer? #1613

Open timoderbeste opened 5 years ago

timoderbeste commented 5 years ago

I am trying to understand the code for WikiSum. I understand that before training we need to download the data and then process it by using the extractive method. However, I am not sure why we are training the wikisum abstractive summarization using a "transformer" model as specified on the README. How do I know that if only the decoder is used or the encoder is also used? And where can I find the implementation for T-DMCA (Transformer Decoder with Memory-Compressed Attention)? Thank you for any help!

zide05 commented 4 years ago

same question !