assertion failed: [When calling zero_state of AttentionWrapper attention_wrapper: Non-matching batch sizes between the memory (encoder output) and the requested batch size. Are you using the BeamSearchDecoder? If so, make sure your encoder output has been tiled to beam_width via tf.contrib.seq2seq.tile_batch, and the batch_size= argument passed to zero_state is batch_size * beam_width.] [Condition x == y did not hold element-wise:] [x (net/decoder1/attention_decoder/rnn/strided_slice:0) = ] [32] [y (net/decoder1/attention_decoder/BahdanauAttention/strided_slice_1:0) = ] [15]
Caused by op 'net/decoder1/attention_decoder/rnn/AttentionWrapperZeroState/assert_equal/Assert/Assert', defined at:
File "C:\Program Files (x86)\Microsoft Visual Studio 14.0\Common7\IDE\Extensions\Microsoft\Python Tools for Visual Studio\2.2\visualstudio_py_launcher.py", line 78, in
File "C:\Program Files (x86)\Microsoft Visual Studio 14.0\Common7\IDE\Extensions\Microsoft\Python Tools for Visual Studio\2.2\visualstudio_py_debugger.py", line 2483, in debug
exec_file(file, globals_obj)
File "C:\Program Files (x86)\Microsoft Visual Studio 14.0\Common7\IDE\Extensions\Microsoft\Python Tools for Visual Studio\2.2\visualstudio_py_util.py", line 111, in exec_file
exec_code(code, file, global_variables)
File "C:\Program Files (x86)\Microsoft Visual Studio 14.0\Common7\IDE\Extensions\Microsoft\Python Tools for Visual Studio\2.2\visualstudio_py_util.py", line 87, in exec_code
exec(code_obj, global_variables)
File "C:\Users\User\Documents\Visual Studio 2015\Projects\Tachatron\Tachatron\eval.py", line 69, in
eval()
File "C:\Users\User\Documents\Visual Studio 2015\Projects\Tachatron\Tachatron\eval.py", line 27, in eval
g = Graph(is_training=False)
File "C:\Users\User\Documents\Visual Studio 2015\Projects\Tachatron\Tachatron\train.py", line 49, in init
File "C:\Users\User\Documents\Visual Studio 2015\Projects\Tachatron\Tachatron\network.py", line 85, in decode1
dec = attention_decoder(dec, memory, num_units=hp.embed_size) # (N, T', E)
File "C:\Users\User\Documents\Visual Studio 2015\Projects\Tachatron\Tachatron\modules.py", line 251, in attention_decoder
dtype=tf.float32) #( N, T', 16)
File "C:\Users\User\AppData\Local\Programs\Python\Python35\lib\site-packages\tensorflow\python\ops\rnn.py", line 548, in dynamic_rnn
state = cell.zero_state(batch_size, dtype)
File "C:\Users\User\AppData\Local\Programs\Python\Python35\lib\site-packages\tensorflow\contrib\seq2seq\python\ops\attention_wrapper.py", line 659, in zero_state
message=error_message)]):
File "C:\Users\User\AppData\Local\Programs\Python\Python35\lib\site-packages\tensorflow\python\ops\check_ops.py", line 318, in assert_equal
File "C:\Users\User\AppData\Local\Programs\Python\Python35\lib\site-packages\tensorflow\python\framework\ops.py", line 1269, in init
self._traceback = _extract_stack()
InvalidArgumentError (see above for traceback): assertion failed: [When calling zero_state of AttentionWrapper attention_wrapper: Non-matching batch sizes between the memory (encoder output) and the requested batch size. Are you using the BeamSearchDecoder? If so, make sure your encoder output has been tiled to beam_width via tf.contrib.seq2seq.tile_batch, and the batch_size= argument passed to zero_state is batch_size * beam_width.] [Condition x == y did not hold element-wise:] [x (net/decoder1/attention_decoder/rnn/strided_slice:0) = ] [32] [y (net/decoder1/attention_decoder/BahdanauAttention/strided_slice_1:0) = ] [15]
assertion failed: [When calling zero_state of AttentionWrapper attention_wrapper: Non-matching batch sizes between the memory (encoder output) and the requested batch size. Are you using the BeamSearchDecoder? If so, make sure your encoder output has been tiled to beam_width via tf.contrib.seq2seq.tile_batch, and the batch_size= argument passed to zero_state is batch_size * beam_width.] [Condition x == y did not hold element-wise:] [x (net/decoder1/attention_decoder/rnn/strided_slice:0) = ] [32] [y (net/decoder1/attention_decoder/BahdanauAttention/strided_slice_1:0) = ] [15]
Caused by op 'net/decoder1/attention_decoder/rnn/AttentionWrapperZeroState/assert_equal/Assert/Assert', defined at:
File "C:\Program Files (x86)\Microsoft Visual Studio 14.0\Common7\IDE\Extensions\Microsoft\Python Tools for Visual Studio\2.2\visualstudio_py_launcher.py", line 78, in
File "C:\Program Files (x86)\Microsoft Visual Studio 14.0\Common7\IDE\Extensions\Microsoft\Python Tools for Visual Studio\2.2\visualstudio_py_debugger.py", line 2483, in debug
File "C:\Program Files (x86)\Microsoft Visual Studio 14.0\Common7\IDE\Extensions\Microsoft\Python Tools for Visual Studio\2.2\visualstudio_py_util.py", line 111, in exec_file
File "C:\Program Files (x86)\Microsoft Visual Studio 14.0\Common7\IDE\Extensions\Microsoft\Python Tools for Visual Studio\2.2\visualstudio_py_util.py", line 87, in exec_code
File "C:\Users\User\Documents\Visual Studio 2015\Projects\Tachatron\Tachatron\eval.py", line 69, in
File "C:\Users\User\Documents\Visual Studio 2015\Projects\Tachatron\Tachatron\eval.py", line 27, in eval
File "C:\Users\User\Documents\Visual Studio 2015\Projects\Tachatron\Tachatron\train.py", line 49, in init
File "C:\Users\User\Documents\Visual Studio 2015\Projects\Tachatron\Tachatron\network.py", line 85, in decode1
File "C:\Users\User\Documents\Visual Studio 2015\Projects\Tachatron\Tachatron\modules.py", line 251, in attention_decoder
File "C:\Users\User\AppData\Local\Programs\Python\Python35\lib\site-packages\tensorflow\python\ops\rnn.py", line 548, in dynamic_rnn
File "C:\Users\User\AppData\Local\Programs\Python\Python35\lib\site-packages\tensorflow\contrib\seq2seq\python\ops\attention_wrapper.py", line 659, in zero_state
File "C:\Users\User\AppData\Local\Programs\Python\Python35\lib\site-packages\tensorflow\python\ops\check_ops.py", line 318, in assert_equal
File "C:\Users\User\AppData\Local\Programs\Python\Python35\lib\site-packages\tensorflow\python\util\tf_should_use.py", line 170, in wrapped
File "C:\Users\User\AppData\Local\Programs\Python\Python35\lib\site-packages\tensorflow\python\ops\control_flow_ops.py", line 124, in Assert
File "C:\Users\User\AppData\Local\Programs\Python\Python35\lib\site-packages\tensorflow\python\ops\gen_logging_ops.py", line 37, in _assert
File "C:\Users\User\AppData\Local\Programs\Python\Python35\lib\site-packages\tensorflow\python\framework\op_def_library.py", line 767, in apply_op
File "C:\Users\User\AppData\Local\Programs\Python\Python35\lib\site-packages\tensorflow\python\framework\ops.py", line 2506, in create_op
File "C:\Users\User\AppData\Local\Programs\Python\Python35\lib\site-packages\tensorflow\python\framework\ops.py", line 1269, in init
InvalidArgumentError (see above for traceback): assertion failed: [When calling zero_state of AttentionWrapper attention_wrapper: Non-matching batch sizes between the memory (encoder output) and the requested batch size. Are you using the BeamSearchDecoder? If so, make sure your encoder output has been tiled to beam_width via tf.contrib.seq2seq.tile_batch, and the batch_size= argument passed to zero_state is batch_size * beam_width.] [Condition x == y did not hold element-wise:] [x (net/decoder1/attention_decoder/rnn/strided_slice:0) = ] [32] [y (net/decoder1/attention_decoder/BahdanauAttention/strided_slice_1:0) = ] [15]