Closed hh23333 closed 5 years ago
Thanks for your question!
The encoder_state is the the network state vector outputted by tf.nn.static_rnn, which is then used in rnn_decoder as the initial_state. The hidden outputs of rnn is stored in encoder_outputs.
Thanks for you reply!
The out put of tf.nn.static_rnn: Returns: A pair (outputs, state) where:
In your paper the network state st and sn is the newgate of GRU, not the last timesteps GRU hidden outputs. So my question is, is there a difference between the presentation of paper and the implementation of the code.
I see. I didn't look at the details of GRU in Tensorflow when I implemented our model. If Tensorflow implements the GRU as the way you described it, then the notation in the paper is a little bit misleading. Thanks for pointing this out.
Hi QingyaoAi, Thanks for releasing code, it seems that there is an ambiguity in regard to the meaning of St(new state) in your paper and code. In the paper, St in formula(4) means new_gate in GRU. But in your code it seems that St is the last timestep hidden outputs:
encoder_outputs, encoder_state = tf.nn.static_rnn(enc_cell, encoder_embed, dtype=dtype)
line 664 of RankLSTM_model.pyCan you explain it to me? Thanks!