Closed matirojasg closed 3 years ago
Hi @matirojasg,
Thank you for reporting this issue.
It should be resolved now on both models.
Also all the configs were updated and added support for both fast and legacy tokenizers and for both frameworks, PyTorch and Tensorflow.
Regards
Hi, there's something strange with the model using transformers library:
So it returns a wrong value for model_max_length - for another model like BERTurk it returns the correct value.
The easiest way would be to extend the tokenizer_config.json and add a "max_len": 512 option :)
Reference