-
Hi, @nicolas-ivanov
I run the training code.Maybe it contains some errors in it.I get the error like below:
`ValueError: Shape mismatch: x has 64 rows but z has 24 rows
Apply node that caused the …
-
First of all, congratulations for this fantastic work!
Regarding the encoder-decoder architecture, have you ever considered using a seqquence-2-sequence model? In a seq2seq model, the decoder's inp…
-
当运行a1_seq2seq_attention_train.py文件时 遇见下面的错误。希望得到您的帮助。
ValueError: Variable W_initial_state1 already exists, disallowed. Did you mean to set reuse=True in VarScope? Originally defined at:
File "/…
-
Thanks for sharing! Just found out `Attention.get_att_weight` is calculating attention in a for-loop? this looks rather slow isn't it?
`4-2.Seq2Seq(Attention)/Seq2Seq(Attention).ipynb`
```pyth…
-
-
Is there a way to pass extra feature along with the existing word tokens as input and feed it to the encoder RNN?
Lets consider the NMT problem , say I have 2 more feature columns for the correspon…
-
Hi, all
I have confusion about this:
decoder_hidden = encoder_hidden[:decoder_test.n_layers] # Use last (forward) hidden state from encoder,
should this be
decoder_hidden = encoder_hidden…
-
Problem Description
This notebook demonstrates how to instruction tune SeqSeq models using huggingface transformers. Instruction tuning is a machine learning paradigm where a model is trained to foll…
-
Hello, is the dataset used in seq2seq + attention in the paper multiple rounds or single theory?
Is the multi-round divided into a single round?
-
## Title & Topic
- RNN에서의 alignment 의미와 seq-to-seq 모델
## Upload schedule
- 2019-03-31
## Reference
- 본문 내에 있음