suriyadeepan / practical_seq2seq

A simple, minimal wrapper for tensorflow's seq2seq module, for experimenting with datasets rapidly
http://suriyadeepan.github.io/2016-12-31-practical-seq2seq/
GNU General Public License v3.0
570 stars 270 forks source link

Deep explaination regards #65

Open karthika-devi opened 6 years ago

karthika-devi commented 6 years ago

Hi sir, As I am new to seq2seq model.I understand how you construct data and why you are using rnn with lstm.But I want to know whether seq2seq can be used for conversation. please explain the functionality of this tf.contrib.legacy_seq2seq.embedding_rnn_seq2seq,tf.contrib.rnn.MultiRNNCell,tf.contrib.rnn.BasicLSTMCell,tf.contrib.rnn.DropoutWrapper

I can understand theoritically what is multiRNN,lstm,dropout,seq2seq but how input get interacted with all those things and why we construcrt this is needed for me to proceed my work

After training get finished I got same word repeated as a result eg: Question: [big unk so far how about you]; answer : [classic classic classic classic classic classic classic classic classic classic classic classic classic classic classic classic classic tie tie tie tie wendell wendell wendell wendell] please guide me why I got this answer and where I want to improve

Thank you

karanpande commented 6 years ago

Each of them functions you mentioned works on the input in a different manners, for example: dropout will prevent the model from overfitting (it just drops nodes). LSTM will help the model not to make grammatical mistakes. Rnn and lstm work in tandem. And so on

With regards to you output- There could be multiple reasons why your output is the way it is. 1: May be the vocabulary size is small try to increase it 2: try training the model on a different dataset 3: increase the number of epochs

Sent from my iPhone

On Jun 24, 2018, at 2:05 PM, karthika-devi notifications@github.com<mailto:notifications@github.com> wrote:

Hi sir, As I am new to seq2seq model.I understand how you construct data and why you are using rnn with lstm.But I want to know whether seq2seq can be used for conversation. please explain the functionality of this tf.contrib.legacy_seq2seq.embedding_rnn_seq2seq,tf.contrib.rnn.MultiRNNCell,tf.contrib.rnn.BasicLSTMCell,tf.contrib.rnn.DropoutWrapper

I can understand theoritically what is multiRNN,lstm,dropout,seq2seq but how input get interacted with all those things and why we construcrt this is needed for me to proceed my work

After training get finished I got same word repeated as a result eg: Question: [big unk so far how about you]; answer : [classic classic classic classic classic classic classic classic classic classic classic classic classic classic classic classic classic tie tie tie tie wendell wendell wendell wendell] please guide me why I got this answer and where I want to improve

Thank you

— You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHubhttps://github.com/suriyadeepan/practical_seq2seq/issues/65, or mute the threadhttps://github.com/notifications/unsubscribe-auth/AKaB9BBQuR3eTx4BJ5eseDSqOzchmSxZks5t_9TrgaJpZM4U1PmG.