Closed Luckick closed 2 years ago
Hi,
Thanks for your interest in our repository. Unfortunately, the current version does not support seq2seq models yet. But technically the approach could be adapted to seq2seq models. I'd be happy to help if you have any issues implementing it!
Does the pruning technique work only for encoder layers or for both encoder and decoder layers?
It should work for both but would require placing the masking variables on the decoder side and cross-attentions properly.
Thank you for the clarification and help!
Hi, Does CoFiPruning work on Encoder-Decoder Architectures for Seq2seq tasks such as translation? Thanks!