Open learningneo opened 6 years ago
Hi,
So even after changing the parameters target_vocab
to 10000 and emb_size
to 200 and batch_size
to 24, still facing the OOM error. Please let me know if I am doing something wrong here or if there is any other parameters/parameter values which I need to try with, given my infrastructure.
Were you able to solve this? We too are not able to run the released code end to end.
Nah.. I tried to run it on a huge server with CPU - even there, got OOM error. So my only guess is the author is running on large multi GPU machines, but can't be sure at all. Waiting for the author/team to comment further, when ever possible.
Sorry for the late response!
It's very confused to me why tensorflow tries to allocate a tensor with shape [32,1226,20003]. Because we limit the length of input sequences as 100 in DataLoader.py (line 93-98).
if max_text_len > self.man_text_len: # self.man_text_len = 100
text = text[:self.man_text_len]
field = field[:self.man_text_len]
pos = pos[:self.man_text_len]
rpos = rpos[:self.man_text_len]
text_len = min(text_len, self.man_text_len)
So the input sequence whose length is 1226 seems impossible (should by less than 100) in our configuration.
Actually, we never met OOM problem while running on a single GTX1080Ti GPU (11G RAM). It takes about 4.4G RAM to run our code with batch size 32.
any update of the error?
whati is the equivalent of of print 'stty' in windows?
I also encountered this problem, but when I changed the batch_size to 16, it no longer appeared.
Hello, Thanks for providing the code. The requirement.txt is missing, but I was largely able to set the code and run it. I am facing runtime OOM error:
I am using a TITAN X GPU with 12 GiB of RAM. What parameters should be decreased/controlled to get the prototype running? Any help will be much appreciated.