-
# DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter
BERTを知識蒸留することにより、軽量化・高速化を図った研究。モデルの性能を97%保ちつつ、モデルサイズを40%、計算速度を60%落とすことに成功。知識蒸留を行う際には、教師モデルの出力分布を教師データとしたsoft target lo…
-
### Feature request
When attempting to use this reranker model [mxbai-rerank-large-v1](https://huggingface.co/mixedbread-ai/mxbai-rerank-large-v1) from huggingface with TEI, I got the following error…
-
**Describe the bug**
Getting the following errors:
Attempted request to https://api-inference.huggingface.co/models/distilbert-base-cased-distilled-squad failed: HTTP/1.1 503 Service Unavailable - {…
-
How can I use DistilBERT for multi-label classification for building a fast and deploy-able model?
-
Some weights of the model checkpoint at distilbert-base-uncased were not used when initializing DistilBertForSequenceClassification: ['vocab_projector.bias', 'vocab_layer_norm.bias', 'vocab_transform.…
-
Sorry if this is a noob question, but I'm wondering if there is a straightforward way to use this model for basic sentiment analysis (positive/negative), similar to how "distilbert-base-uncased-finetu…
-
When calling:
model = SentenceTransformer('distilbert-base-nli-stsb-mean-tokens')
I consistently receive timeout errors.
MaxRetryError: HTTPSConnectionPool(host='sbert.net', port=443): Max re…
-
## 집현전 중급반 스터디
- 2022년 6월 26일 일요일 9시
- 장동건님 김제우님 김종은님 이기성님 발표
- 논문 링크: https://arxiv.org/abs/1910.01108
> ### Abstract
> As Transfer Learning from large-scale pre-trained models becomes more pr…
-
Fine tune 3 (or more) popular models and compare performance to DistilBERT for the movie sentiment analysis task.
Some choices:
GPT-3
LaMDA
Turing-NLG
XGen
Llama 2 (7 billion)
Gemini
Pic…
-
Hi All,
I fine tuned "distilbert-base-uncased" for text classification on a custom dataset. The best accuracy on the validation set was 0.9009 as you can see here: https://huggingface.co/anuragrawa…