Closed Nguyendat-bit closed 2 years ago
"tokenizers" likely makes use of a fast variant. And unfortunately, PhoBERTtokenizer currently does not support it atm. See: https://github.com/huggingface/transformers/pull/13788#pullrequestreview-771521131
@Nguyendat-bit The process of merging a fast tokenizer for PhoBERT is in the discussion, as detailed in https://github.com/huggingface/transformers/pull/17254#issuecomment-1133932067. While waiting for this pull request's approval, if you would like to experiment with the fast tokenizer, you might install transformers as follows:
git clone --single-branch --branch fast_tokenizers_BARTpho_PhoBERT_BERTweet https://github.com/datquocnguyen/transformers.git
cd transformers
pip3 install -e .
@datquocnguyen Great, thank you so much
Hi, I'm getting this error now, is there any way to fix it