Closed prattcmp closed 9 months ago
Right now the tokenizer is private, so it can't be used to count the number of tokens in a passage or to chunk passages to a specific token limit.
This PR moves towards resolving that.
Thank you.
:tada: This PR is included in version 1.10.0 :tada:
The release is available on:
v1.10.0
Your semantic-release bot :package::rocket:
Right now the tokenizer is private, so it can't be used to count the number of tokens in a passage or to chunk passages to a specific token limit.
This PR moves towards resolving that.