Closed wangcongcong123 closed 4 years ago
Yes. I too have faced this issue. Couldn't get any lead for solving this issue from online.
On Wed, Aug 19, 2020, 6:35 PM Congcong Wang notifications@github.com wrote:
I tried ![this tagging tool](from https://github.com/aritter/twitter_nlp) as specifed in this repository. After testing it on Mac, Windows, and Linux, the following still raises.
"module html.parser not found"
I searched solutions to this issue online, none can properly help fix this issue.
does anyone encounter the same issue?
— You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub https://github.com/viczong/extract_COVID19_events_from_Twitter/issues/13, or unsubscribe https://github.com/notifications/unsubscribe-auth/AGM64F25LZPJB7T4TCIRSCTSBPE2NANCNFSM4QE5MZTA .
Oh really? I have been using it for a very long time. I will take a look at this.
A following-up question related to this is that why the tagging process is needed now that the char-chunks offsets are already in the datatset?
The input of the model should be the tokenized text.
@wangcongcong123 Can you provide more information for this issue (e.g., screenshot for this matter, py version, linux environment, etc.)? It looks all good at my side.
Below is the full error message. I use python 2.7
Traceback (most recent call last):
File "python/ner/extractEntities2_json.py", line 26, in
My Linux environment is: Linux xxx 5.4.0-42-generic #46~18.04.1-Ubuntu SMP Fri Jul 10 07:21:24 UTC 2020 x86_64 x86_64 x86_64 GNU/Linux
So it seems it is to the version issue of "twokenize". Actually there is a twokenize.py in tagging's repo. Maybe you could try to use that instead of the pre-installed version. And the tagging tool seems not using "html.parser".
I solved this my making sure I ran the bash build.sh in the tagging repo and not pip installing twokenize separately, as it is some other package built for python 3. Make sure twokenize is using the tagging repo's tokenizing code and not some other pip package.
Can you please explain.. I had run the bash in tagging repo. How to add the ner to our working code.
On Sat, Aug 29, 2020, 7:29 PM Max Weinzierl notifications@github.com wrote:
I solved this my making sure I ran the bash build.sh in the tagging repo and not pip installing twokenize separately, as it is some other package built for python 3. Make sure twokenize is using the tagging repo's tokenizing code and not some other pip package.
— You are receiving this because you commented. Reply to this email directly, view it on GitHub https://github.com/viczong/extract_COVID19_events_from_Twitter/issues/13#issuecomment-683294132, or unsubscribe https://github.com/notifications/unsubscribe-auth/AGM64F6UC3TXLU6KKEEBHGDSDECSTANCNFSM4QE5MZTA .
I tried this tagging tool as specified in this repository. After testing it on Mac, Windows, and Linux, the following issue still raises.
"module html.parser not found"
I searched solutions to this issue online, none can properly help fix this issue.
does anyone encounter the same issue?
A following-up question related to this is that why the tagging process is needed now that the char-chunks offsets are already in the datatset?