huggingface / tflite-android-transformers

DistilBERT / GPT-2 for on-device inference thanks to TensorFlow Lite with Android demo apps
Apache License 2.0
391 stars 81 forks source link

Discrepencies for a fine-tuned DistilBERT model #10

Closed sayakpaul closed 4 years ago

sayakpaul commented 4 years ago

I am trying to develop materials for mobile developers to provide a way for them to compare BERT-based models that are specifically developed for mobile deployments. Currently, I have chosen MobileBERT and DistilBERT for that (repository).

Here's what I have done so far -

Surprisingly, the TensorFlow Lite models achieve a random performance (~50% accuracy) on the development set. This is sharply in contrast with the original fine-tuned model performance (accuracy) which is about ~90%.

I am wondering if I am missing out on something. Any pointers would be really helpful.

Cc: @khanhlvg