issues
search
alexa
/
bort
Repository for the paper "Optimal Subarchitecture Extraction for BERT"
Apache License 2.0
469
stars
39
forks
source link
issues
Newest
Newest
Most commented
Recently updated
Oldest
Least commented
Least recently updated
Bump numpy from 1.16.2 to 1.22.0
#12
dependabot[bot]
opened
2 years ago
0
The exact English pretraining data and Chinese pretraining data that are exact same to the BERT paper's pretraining data.
#11
guotong1988
opened
3 years ago
1
Pre-training-Using-Knowledge-Distillation is better than Pre-training-Only for downstream tasks?
#10
guotong1988
closed
3 years ago
2
Mask-Filling with pretrained BORT
#9
patrickvonplaten
opened
3 years ago
3
how to train model on another language?
#8
Archelunch
closed
3 years ago
1
Create pretraining data with multiprocessing not Implemented
#7
7AM7
closed
3 years ago
1
bort pretrain
#6
nicexw
closed
3 years ago
5
Can't download model again!
#5
killua-zyk
closed
3 years ago
1
Huggingface support
#4
sbsky
closed
3 years ago
2
I couldn't understand the configuration of the model. please can someone clarify?
#3
preethamgali
closed
3 years ago
1
Accuracy during fine-tuning is very low (only 0.68)
#2
waugustus
closed
3 years ago
1
Can't download model.
#1
hardfish82
closed
3 years ago
1