Closed gwarmstrong closed 1 month ago
Fixes a bug in tokenize_batch that occurs when there are more tokens than the specified max sequence length.
tokenize_batch
# Add a code snippet demonstrating how to use this
Pre checks:
max_steps=-1
validation
Needs #354
edit: needs https://github.com/NVIDIA/NeMo-Aligner/pull/355
What does this PR do ?
Fixes a bug in
tokenize_batch
that occurs when there are more tokens than the specified max sequence length.Changelog
Usage
Before your PR is "Ready for review"
Pre checks:
Checklist when contributing a new algorithm
max_steps=-1
andvalidation
?Additional Information