Closed ARDivekar closed 2 years ago
Yes. it should be top-k.
After reviewing the whole process, it can be only drawn words from the entire corpus. Although we can leverage CBOW to predict the target word (i.e. newly inserted word), we are lack of trained neural networks output layer from those pre-trained models. Therefore, we cannot use this approach. Here is the flow of detail implementation in gensim
This is the line: https://github.com/makcedward/nlpaug/blob/master/nlpaug/augmenter/word/word_embs.py#L125
I think it should be
self.model.predict()
, but I am not sure.