Closed lewtun closed 6 months ago
Could there edge cases in which the eos token is not defined in the tokenizer?
Could there edge cases in which the eos token is not defined in the tokenizer?
I'm not aware of any LLM tokenizers that don't have an EOS token, but I think in the worst case we'll have tokenizer.eos_token_id=None
which is the previous behaviour in generate()
Closes #109
I'm not sure if there's any reason not to specify the EOS token ID, but I have verified that adding this ensures chat models terminate on the EOS token.