Since the default tokenizer does not segment Chinese text well, I have added an interface that generates a word cloud based on words and their corresponding weights. This allows users to use other tokenizers to segment the text and obtain a hash table containing keys and their frequencies, and then generate a word cloud based on this hash table.
This PR does not break any public interfaces.
PS: I apologize for the code formatting changes in this PR, as my cargo fmt was automatically invoked.
Since the default tokenizer does not segment Chinese text well, I have added an interface that generates a word cloud based on words and their corresponding weights. This allows users to use other tokenizers to segment the text and obtain a hash table containing keys and their frequencies, and then generate a word cloud based on this hash table.
This PR does not break any public interfaces.
PS: I apologize for the code formatting changes in this PR, as my
cargo fmt
was automatically invoked.