-
1.is there an attention on Intent Detection task in your code? you seem to just make a full-connected layer after Bi-LSTM. That confused me.
2.why the final intent vector is encoder_final_state_h…
-
Hi, I mean without doing something not in the paper [dos santos 2016](http://arxiv.org/abs/1602.03609)
I am mentioning 70% coz it is what the author of this paper reported on using the LSTM+ attenti…
-
1.运行此代码时
![image](https://user-images.githubusercontent.com/31821866/118598731-20c15d80-b7e1-11eb-8fb4-d95cb52a59f2.png)
2.报了以下错误
[2021-05-18 13:49:17,554] [ WARNING] - The _initialize method in …
-
As far as I know, image caption can be generated without the mechanism of attention.
How can i remove the mechanism of attention effectively for training?
can u give me advice pls?
-
I get the following warnings and errors. Sorry if hard to parse/read.
> WARNING:tensorflow:Gradients do not exist for variables ['attention_lstm/kernel:0', 'attention_lstm/recurrent_kernel:0', 'att…
-
I just raised it, as we intensively use RNN (LSTM) for attention-based end-to-end ASR https://github.com/espnet/espnet with chainer and pytorch as a backend, and want to unify these two backends with …
-
I find the output is 1*2 vector in Class CorefTagger,and the final output y is a 1*3 vector in your paper. Are there any differences?
another question: did you test on ".auto_conll" in you paper (…
-
Dear authors,
Thank you for sharing this awesome work. I really like ideas of using all history of words to produce the attention as well as the newly proposed symmetric form for attentions.
Recen…
-
您好,可否提供一下标注文件的生成脚本,比如下面这些已有文件如何生成的
entity_linking/LSTM_CRF/data/lc_quad_entitylink_dev.txt
relation_detection/Attention_BiLSTM/data/Complex_question.test.withpool.nlp.txt
-
Could you add some notes to these models, especially the corresponding paper links? For example, in this paper, Enhancing the locality and breaking the memory bottleneck of transformer on time series …