Closed SaizhuoWang closed 5 years ago
Did you find a solution to this? I am running into the same issue, using a GTX 1080 TI.
@vsub21 Well, it seems that the structure of the network and the batch size have something to do with this issue. I reduced the memory consumption by reducing the complexity of the network(lowering "nr_resnet" and "nr_filters" parameters) and reducing the batch size.
Hi,
yes the code is very memory intensive. It is possible that some refactoring could help with this issue.
Hi there, thanks for your work. I am running main.py on one 1080Ti GPU with memory of 11172MB. And the parameters are all set by default. It seems that the PixelCNN++ model has consumed all the memory and I met this error:
I used pdb to trace the program and it seems that a
u = self.u_stream[i](u, a=u_list.pop())
operation takes about 500MB of memory. And the program ran out of memory after executingu_out, ul_out = self.up_layers[i](u_list[-1], ul_list[-1])
twice, each execution taking about 6000MB of memory. Can you help me with this? I don't know if it is normal with the default parameter setup.