Open p11188536 opened 10 months ago
I want to remove tokens that exceed the max_seq_len. How can I achieve this functionality?
token_in = tokenizer.encode(input) // do your python array slice limit <= max_seq_len here // ask bing/gpt how to return a sub-slice of a python slice/array
I want to remove tokens that exceed the max_seq_len. How can I achieve this functionality?