Closed lostmsu closed 5 years ago
Could a simple way of doing this be to report the model's perplexity on WikiText-103?
It would be easy to compare these numbers to those in the original paper (26.4/17.5 PPL for 345/1542M).
I tried to approximate the perplexity on Wikitext-103. Got about 59 for PrettyBig model. Colab notebook
I am working with OpenAI to produce accurate comparisons. More information will come soon.
Thanks, great to hear!
Hi Connor,
I'd like to see some meaningful comparison with released, and, if possible, unreleased OpenAI's pretrained GPT-2 models.
My concern is that if you used different training techniques, the result may be very far off from what they've got. Including a possibility, that 1.5B model could be worse, than 345M model, that they have released.
P.S. Also pinged you on Twitter about this.