afshinrahimi / mmner

Massively Multilingual Transfer for NER
Apache License 2.0
85 stars 9 forks source link

De-tokenized text #6

Open thangld201 opened 3 weeks ago

thangld201 commented 3 weeks ago

Hi @afshinrahimi, @yuan-li, do you still keep the raw data (un-tokenized one) ? Also, which tokenizer did you use for this dataset ?

I need to work with the raw form of the text, and de-tokenizing is untrivial for many languages (e.g. korean, japanese).

afshinrahimi commented 2 weeks ago

Hi Thang, Have you looked at the data? It was a long time ago. What I remember is that we did not use any special tokenizer. For this reason, we didn't use JP, KO, or other languages for which a simple whitespace tokenizer wouldn't work. You can see the list of languages used in the paper as a footnote:

af, ar, bg, bn, bs, ca, cs, da, de, el, en, es, et, fa, fi, fr, he, hi, hr, hu, id, it, lt, lv, mk, ms, nl, no, pl, pt, ro, ru, sk, sl, sq, sv, ta, tl, tr, uk and vi.

This is the original paper that published wikiann dataset: https://aclanthology.org/P17-1178.pdf. In the paper there is a link to the original dataset that doesn't work anymore but you might be able to use archive.org's time machine https://web.archive.org/ to find it.

Also look at the data (not hosted on huggingface) here: https://www.dropbox.com/s/12h3qqog6q4bjve/panx_dataset.tar, if there is a ja dataset there, we have not tokenized it, I believe.

Apologies I can't help more than this.