Closed ejls closed 9 years ago
My mistake, I removed this commit. The sole difference with the repo I used to get the 10.43 bleu score on wmt15 fi-en is that I used the following reshuffled dataset: basedir = '/data/lisatmp3/simonet/wmt15/data/' config['src_data'] = basedir + 'all.tok.clean.shuf2.seg1.fi-en.fi' config['trg_data'] = basedir + 'all.tok.clean.shuf2.fi-en.en'
Okay @ejls , 20% speed up is really amazing, thanks for the effort
Just a note, that the instead of using a custom LookupFeedbackWMT15
you guys could have a special token to be used as initial output. That seems to be even better, since the feedback at the first step would be trainable instead of being zeros.
@rizar thank you for the pointer, actually with the recent changes, we are using separate indices for <S>
and </S>
during training and we do not need LookupFeedbackWMT15
anymore as you stated. LookupFeedbackWMT15
was there just to match costs etc with Ghog models. I will remove it from the MT example as-well.
The --subtensor-fix parameter gives a speed up of 18-20% on training.
To check the correctness of the code I ran both algorithms in parallel :
Disabling theano optimizations (and dropout/noise) the updates are exactly the sames, but with optimizations enabled I get some 1e-9 differences.