-
**Describe the bug**
Specifying a filter in a query made with dense vectors while working with OpenDistroElasticsearchDocumentStore and EmbeddingRetriever throws a `KeyError: 'script_score'.`
**…
-
# Is there any format converting on corpus (like step 1 of general distillation in the original TinyBERT repo)?
${BERT_BASE_DIR}$ includes the BERT-base teacher model.
python pregenerate_traini…
-
# How to optimize an already fine-tuned model from Hugging Face?
Congratulations on the work, it looks amazing 😊
## Details
If there is an already fine-tuned model from Hugging Face for, let's …
-
# 🚀 Feature request
It would be great if the model cards for models would include the model size (i.e., the number of parameters) and then the model hub will allow searching for models by size.
…
-
python 3.6
pytorch 1.7.1
Is the model name changed?
Thanks!
```
Traceback (most recent call last):
File "/home/zhutian/.conda/envs/infospace/lib/python3.6/site-packages/transformers/configur…
-
**Question**
Can you please tell me what and how an input is passed to the ranker model?. searched on web but there is no result related to that. It would be helpful if you provide me proper way of …
-
Hi,
Whenever I try to run this code :
from bertopic import BERTopic
topic_model = BERTopic(verbose=True, embedding_model="paraphrase-TinyBERT-L6-v2", min_topic_size=25)
topics, _ = topic_mod…
-
Tinybert(4L,"hidden_size": 312,"intermediate_size": 1200,)模型随机初始化,结果相差bert-base 9个点(acc83%降到74%)。这种结果正常吗?
hahlw updated
3 years ago
-
看过TinyBERT的论文后,想请教如下几个问题:
(1)预训练的蒸馏阶段,是指在预训练teacher BERT的同时蒸馏 student TinyBERT吗?比如每个epoch蒸馏一次或者其他?因为看到如下示意图,一开始觉得是预训练的同时进行蒸馏。
![image](https://user-images.githubusercontent.com/10300313/70848333-f28…
-
Hello,
Have you done general distillation using the `bert-base-cased` model?
and would you have the `General_TinyBERT_v2(4layer-312dim)` cased model available?
When trying `python3 task_distill…
sv-v5 updated
3 years ago