cmusphinx / g2p-seq2seq

G2P with Tensorflow
Other
669 stars 195 forks source link

What is the input to the encoder network ? #139

Closed mausamsion closed 6 years ago

mausamsion commented 6 years ago

Hi, As you mentioned in the readme that current version uses the attention mechanism from the transformer model. As in the transformer paper they use embeddings of the words as input. Does the current version of g2p-seq2seq also uses the input as vector embeddings of the tokens (in this case characters) or the encoder sees the tokens as they are ?

mausamsion commented 6 years ago

Sorry, just found your comment on this issue.

nurtas-m commented 6 years ago

Hello, @llrootll Yes, as transformer uses embeddings of the words we also uses embeddings in g2p-seq2seq. But, in our case, we uses embeddings of the graphemes and phonemes instead of the embeddings of the words. Because of transformer model trains embedding layer itself you don't need to worry about it.