Closed toddwyl closed 4 years ago
I think if you use the default setting, then each iteration will take 20-30mins on 2080TI or Titian V. The whole training process will take ~one day. For MovingMNIST, please use the code provided on ICLR page.
But if you want to use larger batch size, you should correspondingly change the learning rate. I am not sure why your generation looks so messy, our method is quite robust to the change of hyperpararmeters. I guess there is something wrong with the architecture. Did you build additional decoder or so?
As for the combination of optimizer. Usually I would use something like below. I don't know if your version works.
optimizer = optim.SGD(
[{'params': model.parameters()}, {'params': metric.parameters()}, {'params': Header.parameters()}], lr=0.001,
momentum=0.9, weight_decay=1e-5)
I think if you use the default setting, then each iteration will take 20-30mins on 2080TI or Titian V. The whole training process will take ~one day. For MovingMNIST, please use the code provided on ICLR page.
But if you want to use larger batch size, you should correspondingly change the learning rate. I am not sure why your generation looks so messy, our method is quite robust to the change of hyperpararmeters. I guess there is something wrong with the architecture. Did you build additional decoder or so?
As for the combination of optimizer. Usually I would use something like below. I don't know if your version works.
optimizer = optim.SGD( [{'params': model.parameters()}, {'params': metric.parameters()}, {'params': Header.parameters()}], lr=0.001, momentum=0.9, weight_decay=1e-5)
Thanks, I slove the issue with the separate optimizer and your scheduler. I think maybe I change the num of epoch small one itr that make the learning rate small by scheduler .
Hi ,May I ask how performance you are on MovingMnist? I use code provided on ICLR page, but i get a bad performance
Thank for your answer!
The training process is
too slow
, it cost one hour per itr. When I set thebatch_size
to64
in dataMovingMnist
, it get the bad performance. Is there any training tricks ?My parameters are set as belows:
And I combine the optimizer between
frame_predictor
andencoder
:to:
Is there any wrong ?