Closed samuelazran closed 9 months ago
Thank you for noticing. It was actually nothing to do with the tokenizer but rather the streaming decoder which wasn't catching 2-byte UTF8 characters. It should be fixed with the latest commit:
Thank you for noticing. It was actually nothing to do with the tokenizer but rather the streaming decoder which wasn't catching 2-byte UTF8 characters. It should be fixed with the latest commit:
Thank you so much! Update: I have tested it and it works perfect now.
BTW, thanks for this library, this is in my experience the best inference library!
For example, I prompt it to extract an entity from a text in Hebrew, and it produces: "���� ���" It seems to work ok with language understanding of Hebrew but the decoding of the tokens when generating Hebrew seems buggy only when using Exllamav2 (the same prompt with the same model works perfect with Huggingface Transformers).
What could be the issue? Could it be their custom tokenizer?
I've tested it with the latest exllamav2 (from the commit with the Yi models support addition) with several models including "01-ai/Yi-34B-200K" and "LoneStriker/Yi-34B-200K-6.0bpw-h6-exl2" from Huggingface models.
Attaching here a screenshot.