-
Hello
I am using the mne-bids-pipeline and an EEG-only study. I would like to implement a time-frequency analysis. At the moment, it seems that time-frequency analysis are only generated on the sub…
-
Hanlp may currently be the best Chinese segmentation engine, with much better segmentation results than Jieba.
I am trying to use Hanlp to segment all the Markdown documents locally and then creat…
-
I want to take a few dystopian novels and compare word frequency to see what topics come up most frequently. I imagine I may also want to add in some word type filters so that I don't end up with jus…
-
Instead of a simple word count change it to a word cloud. Not so much a cloud as a count of the frequency of each word in the description. Gather all instances of a word being used into an array with …
-
**Paper**
Character-level Convolutional Networks for Text Classification
**Introduction**
In the realm of text classification, most models have considered the words as the building blocks. This r…
-
While frequency per 10k is a much better measure than raw counts, it's still not completely independent of corpus size as Harald Baayen has argued extensively.
I need to reread his book _Word Frequ…
-
We already do a bag of words, and word counts. Would it be useful to anyone to expose this functionality for other classification uses?
Some other things to consider:
- [ ] N-grams
- [ ] Levens…
Ch4s3 updated
7 years ago
-
Hi,
I am attaching a link containing the book statistics of Malayalam language for all the three cases that we discussed i.e.
- When OCR predictions are added to vocabulary in a Sequential man…
-
I started to use Manuskript 0.11.0 to write a book in Russian.
For such complex language frequency analysis is more difficult.
The current functionality is great, but shows many similar words and ph…
-
I have been using command line symspell and very much want to use a node api where the dictionary is built once and remains in memory. I don't have the programming skills to make this a reality so I a…