-
Hi,
I'm trying to reproduce lighthubert_stage1 and lighthubert_small, but got a big performance gap... Could you please supply more details of your training process (such as lr, scheduler or loss fun…
-
I try to run code in colab but I got some issue like this:
```
contextual document embedding is initiated...
Pandas Apply: 100%
2000/2000 [23:34 512). Running this sequence through the model wil…
-
### System Info
```shell
Working on google colab, and installing optimum with:
!python -m pip install git+https://github.com/huggingface/optimum.git#egg=optimum[onnxruntime]
```
### Who ca…
-
**Describe the bug**
optimum-intel cli converted openvino ir models does not load on the model-server
**To Reproduce**
```
$ uname -r
5.15.0-92-generic
$ cat /etc/os-release
PRETTY_NAME="Ub…
-
## ❓ Questions and Help
### Before asking:
1. search the issues.
2. search the docs.
#### What is your question?
Is the token dictionary for the data2vec 2.0 text model available anywhe…
-
### System Info
- `transformers` version: 4.27.2
- Platform: Linux-6.2.0-76060200-generic-x86_64-with-glibc2.35
- Python version: 3.10.6
- Huggingface_hub version: 0.13.3
- PyTorch version (GPU?)…
-
This sprint is similar to #16292 - but for model **configuration files**, i.e. `configuration_[model_name].py`.
For example, `src/transformers/models/bert/configuration_bert.py`
# The expected ch…
-
Dear authors,
I'd like to cite your paper and let me check if it is to appear at ICASSP.
Thanks for your help!
-
On your results table (https://github.com/HolgerBovbjerg/data2vec-KWS#results), as the kwt parameter increases, the accuracy decreases.
- KWT-1 (full) : 0.9638
- KWT-2 (full) : 0.9498
- KWT-3 (full…
-
hi,
I inference GPT-NeoX and LLaMa-7B with BetterTransformer,but get the same latency with huggingface transformers,
Python: 3.10
PyTorch: 2.0
CUDA: 11.7
transformers: 4.29
op…