Closed LoicDagnas closed 1 year ago
(1) batch size must be known for beam search; (2) target_ids tensor is not necessary for decoding.
(1) it must be the same only when padded_decode is set to true?
(2) yes, in fact it is related to my other issue concerning this model https://github.com/tensorflow/models/issues/10221
Hi @LoicDagnas,
Thank you for opening this issue. Since this issue has been open for a long time, the code/debug information for this issue may not be relevant with the current state of the code base. The TF models official team is constantly improving the framework by fixing bugs and adding new features. We suggest you try the latest TensorFlow version with the latest compatible hardware configuration which could potentially resolve the issue. If you are still facing the issue, please create a new GitHub issue with your latest findings, with all the debugging information which could help us investigate. Please follow the release notes to stay up to date with the latest developments which are happening in the TF models official space.
This issue has been marked stale because it has no recent activity since 7 days. It will be closed if no further activity occurs. Thank you.
This issue was closed due to lack of activity after being marked stale for past 7 days.
1. The entire URL of the file you are using
https://github.com/tensorflow/models/blob/master/official/nlp/nhnet/models.py
2. Describe the bug
While using a Bert2Bert model instance setting the
padded_decode
parameter to True (e.g. for TPU usage), I am forced to specify the batch size in the input when calling the model.3. Steps to reproduce
You can simply run the following code:
you'll get the following stack:
but if you give the following input with the batch size specified:
it will work.
4. Expected behavior
I was expecting that it works in both case i.e. specifying the batch size or not.
5. Additional context
X
6. System information