-
when I intend to train the general chinese tinybert model, I meet some problems that the project doesn't offer the example of the config file, so could anyone offer me a reference of the config file …
-
python train.py --output_dir=‘ouput_dir’ --model_type=bert --model_name_or_path '2nd_General_TinyBERT_4L_312D/' --mlm --do_train --do_eval --eval_data_file=data/restaurant/train.txt --per_gpu_train_ba…
-
Hi,
Thank you so much for open sourcing this toolkit! This is very helpful. I tried to finetune on several downstream tasks and it works great.
May I ask if you ever tried to distill a random-in…
-
Hello!
Do you have any plans for training larger transformers models, something from the last architectures (Reformer specifically) or BERT?
Or maybe you have plans in the opposite direction: t…
-
Sequence-Level Knowledge Distillation, Yoon Kim, Alexander M. Rush, 2016
https://arxiv.org/pdf/1606.07947.pdf
-
I'm trying to follow the example but following either the pip (or docker routes) I get:
```
nboost --uhost localhost
--uport 9200 …
-
Hello,
how to fix this
> CANNOT LINK EXECUTABLE
errors? Not running on Kirin 980 nor 990.
> 1: --- Network Test (LeNet)
> 1: CANNOT LINK EXECUTABLE "/data/local/tmp/uldra/lenet": cannot l…
-
比如TinyBert ALbert?
-
This issue is not specific to Haystack, but I just wanted to know whether this is something that I can achieve using Haystack.
I wish to train two domain-specific models:
Domain 1: Constitution …
-
Hi,
Let's start a discussion here about the roadmap towards 0.10 and 1.0. We are looking for:
- New features that are useful to your research
- Improvements and patches to existing features
If…