-
## 概要
WIKIPEDIAのような固定フォーマットの表からその表全体の記述を行うモデルの提案を行なっている.
そのモデルは表の内容と表のフィールドの情報の両方をそれぞれ入力とするエンコーダ,普通のアテンションと表の構造に関するアテンションを組み合わせたデコーダなどによって構成されている.
WIKIBIOデータセットで精度の向上が見られた
![](https://d2mxuefqeaa7…
-
### Metadata
Authors: Prajit Ramachandran, Peter J. Liu and Quoc V. Le
Organization: Google Brain
Conference: EMNLP 2017
Link: https://goo.gl/n2cKG9
-
I found that in the paper, the formula of MLP attention is usually desribed as below:
![image](https://user-images.githubusercontent.com/16586180/39976766-fd23c30e-5767-11e8-9a16-9d0238512c82.png)
…
-
Since we can handle generic graph-networks now, it's time to implement LSTM/RNN. Milestones:
- [ ] Add a basic LSTM cell (equivalent to TF's BasicLSTMCell)
- [ ] Modify network class to handle sequent…
nyanp updated
7 years ago
-
Hello, I'm trying out your code and I'm running the Test and Visualize attention results codes and I'm getting this the error below. Could it be that the sample model is already outdated? Thank you.
…
-
-
我使用ALBERT和孪生网络来训练一个主观问题评分模型,训练策略参考的你的代码,孪生网络由双向LSTM和全连接层组成。在训练中,我发现准确率没有提高,一直保持不变。我感觉像是权重没有更新,可能是因为梯度太小导致了权重变化不大。或者,训练策略可能存在问题,但我不确定具体原因。下面是我训练期时的准确率:
![training](https://github.com/dragen1860/MAML-P…
-
Thank's for sharing this code. I was wondering if exist a way to use different format of jointand number of joints in training, what would be the changes that we should made in this project.
Thank's …
-
Where is the test dataset BLUE score?
-
### Abstract
- propose additional "Attentive Recurrent Network(ARN)" to Transformer encoder to leverage the strengths of both attention and recurrent networks
- WMT14 EnDe and WMT17 ZhEn demonstra…