-
from spacy import displacy
from tabulate import tabulate
import zh_core_web_sm
nlp = zh_core_web_sm.load()
when I try this, it gives me the following error:
ValueError: Can't read file: D:\Anac…
-
![image](https://user-images.githubusercontent.com/50871412/119260850-4f876b80-bc07-11eb-8894-124302600643.png)
![image](https://user-images.githubusercontent.com/50871412/119260875-675eef80-bc07-11e…
-
Right now the lessons are not very NLP-ish. We had originally talked about having some lessons that illustrate maxent models _in NLP_, on tiny examples.
Below is a quote from my email of 9/10/12 wh…
-
My corpus contains 300 paragraphs, and the speed is slow. More than 30 mins.
Could you please introduce sumy's performance ? And which stage will make it slow when corpus is large.
Thanks!
-
Hello Maarten, there is one thing I would like to mention when using BERTopic to analyze Chinese and Japanese texts. If we run the following code to analyze Chinese or Japanese:
from bertopic impo…
-
Thank you for creating the tool for public use!
I found that the tokenizer does not work well in some occasions. Is there any way to give a delimited input to your POS and dependency parser directly …
-
Goal: better feature/model discoverability for the GluonNLP website
### Side Navigation Bar
* Installation
* Models
* Tutorials
* Demos (in the future)
* API Documentations
* Community…
-
Is there any glitch in the installation?
I'm using Ubuntu.
After
```shell
git clone https://github.com/IDEA-CCNL/Fengshenbang-LM.git
cd Fengshenbang-LM
pip install --editable ./
```
I copi…
-
https://github.com/huggingface/nlp/blob/7d1526dfeeb29248d832f1073192dbf03ad642da/metrics/bleu/bleu.py#L76 assumes the inputs are tokenized by the user. This is bad practice because the user's tokeniz…
-
/chat: Will LLM do word segmentation for Chinese? Or do they simply read each Chinese character and run the process?