Hello,
This is not an issue but rather a question -
Where could I get all the datasets you reported to in the paper ?
Do you think that training on ALL datasets together would improve the results ?
What about training for various languages - do you think a model containing text for mixed languages would behave better or worse than models handling each language separately ?
And another question regarding phrases - the google's word2vec pretrained vectors include also phrases - were they taken into account as well ?
Hello, This is not an issue but rather a question - Where could I get all the datasets you reported to in the paper ? Do you think that training on ALL datasets together would improve the results ? What about training for various languages - do you think a model containing text for mixed languages would behave better or worse than models handling each language separately ?
And another question regarding phrases - the google's word2vec pretrained vectors include also phrases - were they taken into account as well ?