-
https://aclanthology.org/2021.wnut-1.31.pdf
学会 : ACL2021
-
Hi there!
I looked through the corpuses and found that sometimes they are not 100% downloaded. Not sure, if the issue is with the downloading scripts. Below are some examples grepped from bookcorp…
-
Hi Deepspeed team,
The BERT perf results from blog [Microsoft DeepSpeed achieves the fastest BERT training time](https://www.deepspeed.ai/news/2020/05/27/fastest-bert-training.html) are very impres…
-
# My question
Why does catastrophic forgetting occur when I perform continued pre-training on Llama 3? I used open source data from BookCorpus, iterated 100,000 steps, and then after testing the trai…
-
Hu et al. (2019)'s paper is summarised in #35 . Their code is [here](https://github.com/iris2hu/diachronic-sense-modeling).
@kasparvonbeelen started implementing their method in #18
Currently n…
-
## Description
- I want to train BERT model on GPU, but have some problems. My configuration:
* Software environment: Python: 3.7.7, Cuda: 10.2
* Install MXNet: `pip install mxnet-cu102` , ve…
-
- https://arxiv.org/abs/1911.05507v1
- 2019
我々は、過去の記憶を圧縮して長距離シーケンス学習を行う気配りシーケンスモデル「Compressive Transformer」を発表する。
Compressive Transformerは、WikiText-103およびEnwik8ベンチマークにおいて、それぞれ17.1 pplおよび0.97 bpcと…
e4exp updated
3 years ago
-
I think there is a bug in standard sentence tokenizer `sent_tokenize`. The problem is, that it is not splitting text into sentences under certain case. Here is this case, where the tokenizer fails to …
-
### 論文へのリンク
[[arXiv:2004.02984] MobileBERT: a Compact Task-Agnostic BERT for Resource-Limited Devices](https://arxiv.org/abs/2004.02984)
### 著者・所属機関
Zhiqing Sun, Hongkun Yu, Xiaodan Song, Ren…
-
BERT for Joint Intent Classification and Slot Filling
Qian Chen, Zhu Zhuo, Wen Wang
4 pages, 1 figure
https://arxiv.org/abs/1902.10909
## 概要
新しい言語表現モデルであるBERTを利用して,インテント分類及びスロットフィリングを行った.
…