-
I thought we had addressed this already, but maybe this is part of #719.
### for tokens
Define two sets of tokens, simple unigrams and space-separated bigrams:
```r
(toks
-
Not as much an issue as a question. Since you already use data.table in your package, why not go all the way and use it more effectively in your functions?
I have tried doing a txt_freq using data.…
-
I'm having an issue where Unicode encoding characters are ending up in my Hug response. When I save the dictionary to a file using JSON.dumps it doesn't have this issue. I spent a day trying to change…
-
Will need to index the tokens/sentences first
-
### Summary
- What does this package do? (explain in 50 words or less):
The aim of this package is provide some simple functions in R to explore changes in word frequencies over time in a spec…
-
Hi there, I'm using this awesome lib in a chatbot project called [hubot-natural](https://github.com/RocketChat/hubot-natural) and I'm having trouble to use the PoS Tagger feature to recognize Brazilia…
-
@kbenoit A few points prior to describing my feature request:
1. The `pkgdown` documentation is somewhat behind (`kwic`, for example, still notes the [`join` option](http://quanteda.io/reference/kw…
-
Based on feedback from CS410 instructor Ismini Lourentzou I am reporting the following issue regarding the interaction of Ngrams and Stemmers in MeTA.
Briefly: should they not be ***_independent_*…
-
Implement the character n-gram feature for ODIN data. Use the `character-n-gram-size` parameter in the config file and the character_ngrams() function in analyzers.py to compute the n-grams. Use the m…
-
I found this weird thing. dfm_lookup is crashing with the above error in certain cases when 'exclusive=TRUE.' Here's a minimum working example. I don't know why "featured_story_content_h2" is importa…