nii-yamagishilab / multi-speaker-tacotron

VCTK multi-speaker tacotron for ICASSP 2020
BSD 3-Clause "New" or "Revised" License
265 stars 41 forks source link

My Additive-Attention is not good #9

Closed BabaiLi closed 3 years ago

BabaiLi commented 3 years ago

The Additive-Attention has always been bad. According your paper, this Attention helps Forward Attention to align, so is it normal that always bad? My Forward Attention is well aligned.

I use all of the VCTK, and the batch_size is 32. The figure below shows the number of epochs 37, 47, and 48.

53000 67000 70000

BabaiLi commented 3 years ago

I got the answer in "Investigation Of Enhanced Tacotron Text-To-Speech Synthesis Systems With Self-Attention For Pitch Accent Language" this paper, additive-attention is work. Thanks.