Closed chuzheng88 closed 2 years ago
Hi, please take a minute to read the documentation. The shape of the input tensor should be (batch_size, sequence_length, input_dimension)
, so in your case (batch_size, 1000, 2)
.
also error occur
Again, please read the documentation of MultiHeadAttentionChunk
. You're using the Chunk
MHA, which clearly states that the default chunk_size
is 168
. This is why you are getting an error.
why did the error occur ?