A simple but complete full-attention transformer with a set of promising experimental features from various papers
4.42k
stars
377
forks
source link
Is this the same "X-transformer" that being used in "X-Transformer: A Machine Translation Model Enhanced by the Self-Attention Mechanism" paper? #257
Closed
argadewanata closed 1 month ago
I read a paper titled "X-Transformer: A Machine Translation Model Enhanced by the Self-Attention Mechanism" on https://www.mdpi.com/2076-3417/12/9/4502.
Is this the same "X-Transformer"?