bheinzerling / bpemb

Pre-trained subword embeddings in 275 languages, based on Byte-Pair Encoding (BPE)
https://nlp.h-its.org/bpemb
MIT License
1.18k stars 101 forks source link

question on https://nlp.h-its.org #35

Closed jwijffels closed 4 years ago

jwijffels commented 4 years ago

Hi,

Now that issue https://github.com/bheinzerling/bpemb/issues/34 is sorted out. For the R wrapper of sentencepiece at https://github.com/bnosac/sentencepiece, I'm planning to implement a simple wrapper to download the models you have been providing and next put the package on CRAN. Before I do this, I would like to know the intention of that site: https://nlp.h-its.org Namely

Thanks for any input.

bheinzerling commented 4 years ago

Hi again,

I'm responsible for https://nlp.h-its.org/bpemb and it's intended to stay available indefinitely. There are no objections to redistribution or downloading. In fact, BPEmb is already being redistributed or wrapped in some other projects, e.g. by Wolfram and in flair.

jwijffels commented 4 years ago

Thanks for the input. So the license for the sentencepiece models and the Glove embeddings is the same MIT license as the one shown at https://github.com/bheinzerling/bpemb/blob/master/LICENSE?

bheinzerling commented 4 years ago

yes!

jwijffels commented 4 years ago

thanks! you make my day!