Closed VanOvermeire closed 4 years ago
In full bert
case you are using BertForMaskedLM
but for distill bert you are using DistilBertModel
which is not for masked language modelling. Try using DistilBertForMaskedLM
. Check it, it works:
https://colab.research.google.com/drive/1GYt9H9QRUa5clFfAke6KPYl0mi4H1F3H
Well, in hindsight that was obvious. :) Thanks!
Hi,
This is probably me doing something wrong, but I can't get distilbert to give me a sensible prediciton when I mask part of a sentence.
This setup for BERT (based on the examples):
gives the correct answer are for How are you doing?.
But when I try the same with distilbert:
I practically always get some unusedxxx as a result. At first I thought this was because distilbert is a smaller model, but no matter what I try, I keep getting unused, so I am guessing it's something else.
Thanks in advance!