duyvuleo / Transformer-DyNet

An Implementation of Transformer (Attention Is All You Need) in DyNet
MIT License
63 stars 19 forks source link

Does support multi-gpus training ? #4

Open KelleyYin opened 4 years ago

KelleyYin commented 4 years ago

This is a great repo. Can this code support multi-GPU training? I wonder if it can achieve the same performance as tensor2tensor on wmt14-en-de corpus. Thanks.