Open KushalDave opened 5 years ago
I have tried several things but cant get it working. Adding this weight seems to bloat up the memory over 2G and the code crashes.
You could try to change the type of the weights to tf.float16 or something with lower precision to save memory.
Hi Zafarali,
I am trying to use your attention network to learn seq2seq machine translation with attention. My spurce lang output vocab is of size 32,000 and target vocab size 34,000. The following step blows up the RAM usage while making the model (understandably, as its trying to manage a 34K x 34K float matrix):
Here is my model: n_units:128, src_vocab_size:32000,tar_vocab_size:34000,src_max_length:11, tar_max_length:11
Is there any fix for this?