-
Hii!
Recently, I have been working creating user friendly remixing tool with my friend @tripathiarpan20 known as [midiformers](https://github.com/tripathiarpan20/midiformers). I am wanting to furthe…
-
Seems there is no link for the pretrained bert model where can I get it for the inference?
Thanks
-
75 Languages, 1 Model: Parsing Universal Dependencies Universally
Dan Kondratyuk, Milan Straka
Accepted for publication at EMNLP 2019. 17 pages, 6 figures
https://arxiv.org/abs/1904.02099
code: ht…
-
Recent transformers architectures are very famous in NLP: BERT, GPT-2, RoBERTa, XLNET. Did you try to fine-tune them on some NLP task? If so, what was the best Ranger hyper-parameters and learning rat…
-
Hi, I used around 8000000 text sentences while fine tuning the language model but the newly added vocabulary size is only 50000. My data have atleast around 1000000-2000000 tokens to be added. Can, I …
-
## Environment info
- `adapter-transformers` version: 3.2.1
- Platform: Linux-6.2.0-27-generic-x86_64-with-glibc2.37
- Python version: 3.10.9
- PyTorch version (GPU?): 1.13.1 (GPU)
## De…
-
please, can u describe how to use this library?
Which dataset u used and how to prepare it in such a way that it can be feed directly to BERT and generate embeddings.
-
I'm on an macbook pro 2017 with Intel CPU and RX560, using tf-macos alpha3.
I tried to train BERT with TF-macos but the gradient won't get through layer norm layers until I replace `python/keras/laye…
-
您好,麻烦请问一下如果我想只训练bert层后面接的bilstm-crf层,训练中应该怎么冻结bert层的参数呢?
-
### TODO
- [x] KorQuAD로 fine-tuning하기
- [x] #1 에서 제공되는 test 모듈로 테스트하기