-
Related to **Model/Framework(s)**
*(e.g. GNMT/PyTorch or FasterTransformer/All)*
BERT/PyTorch
In Readme.md of https://github.com/NVIDIA/DeepLearningExamples/tree/master/PyTorch/LanguageModel…
-
## 一言でいうと
Bi-directionalのTransformerを事前学習し、QAや文関係推論などのタスクに転移した研究。ELMo(#655)の双方向性と、OpenAIのTransformer転移(#790)をミックスした形になっている。言語モデル学習は双方向でないのでcropした単語を予測する形で学習、文関係学習のため次の文or notを学習させる等の工夫を行っている。
##…
-
Dear NeuroNER authors,
I have downloaded and installed the NeuroNER project as well as the necessary dependencies (tensorflow, python3, etc.). Note that I am using python 3.6.
When I try to appl…
lujea updated
7 years ago
-
Very nice job!!! I tried to use resnet (10, 12, 18) as a backbone of my model, but it didn't improve performance as we expected. In contrast, In our experiments, deeper resnet was mush easier overfitt…
-
-
-
Hello @MhLiao , my scenario is mostly long text line included. Should I retrain the model by changing the parameter "MODEL.ROI_MASK_HEAD.POOLER_RESOLUTION_W" from 64 to wider, say 128 or 196 (with the…
-
Hi 👋, thanks for your great job! And I have some questions about the Text-based MLC to confirm.
1. When using MIMIC-CXR for pre-training, is the label of multi-label classification extracted by CheX…
-
Trying to run the training for the BERT-large topology, unpadded. We set up an nvidia-docker to run the training workload. However, we run into an error for the unpadded run. Here's an excerpt from th…
-
Hello,
I'm running out of memory when using contrastive pretraining using the default config. I'm using 8 GPUs with 40GB each and I ran out of memory even when I decreased the batch size to 1024.