Closed AlanAboudib closed 3 years ago
its only coming with '.' this matches with each alphabets, its probably some special character investigating this
@AlanAboudib you need to use like this
tokenizer = Tokenizer(suffixes=['$'],
prefixes = ['('],
infixes = ['\.'],
exceptions = {"melo": [{"ORTH":"me"}, {"ORTH":"lo"}]}
)
'.' is special character
Although this error shouldn't come even with this usage, it should tokenize every letter separately. I am looking into it
Description
Here is a weird bug
How to Reproduce
Expected Behavior
Tokenization shouldn't produce an error