Open andreaschandra opened 3 years ago
@alamhanz
## setup tokenizer and model
tokenizer = BertTokenizer.from_pretrained("bert-base-cased-finetuned-mrpc")
model = BertForSequenceClassification.from_pretrained("bert-base-cased-finetuned-mrpc")
32GB memory usage
pre-trained BERT failed, loss didn't decrease as expected