Hello,
I have trained the model given in the walk-though for English-German translation. I was trying to access attention weights but I found there is no direct command or way to access it. So I tried using the code given in https://colab.research.google.com/github/tensorflow/tensor2tensor/blob/master/tensor2tensor/notebooks/hello_t2t.ipynb#scrollTo=OJKU36QAfqOC
but found that there are following major differences between the way decoding is done in the walkthrough compared to the above link
1.Walk through has T2T model set to "INFER" whereas the code given in link has mode "EVAL"
2.The beam size and alpha values can be set to a value for decoding in the walkthough case but the same was not done in the link.
I was trying to access the weights as given in the link (using get_att_mats() function) but was not able to set the mode as "INFER" and required beam and alpha values.
So is there any way I can get decode results as given in walk-through with the set beam and alpha values and still get the attention weight values.
Thanks in advance.
Hello, I have trained the model given in the walk-though for English-German translation. I was trying to access attention weights but I found there is no direct command or way to access it. So I tried using the code given in https://colab.research.google.com/github/tensorflow/tensor2tensor/blob/master/tensor2tensor/notebooks/hello_t2t.ipynb#scrollTo=OJKU36QAfqOC but found that there are following major differences between the way decoding is done in the walkthrough compared to the above link 1.Walk through has T2T model set to "INFER" whereas the code given in link has mode "EVAL" 2.The beam size and alpha values can be set to a value for decoding in the walkthough case but the same was not done in the link.
I was trying to access the weights as given in the link (using get_att_mats() function) but was not able to set the mode as "INFER" and required beam and alpha values. So is there any way I can get decode results as given in walk-through with the set beam and alpha values and still get the attention weight values. Thanks in advance.