Hi, the default batch size is 1, did you try the larger batch size. In my experiments, when the batch size >= 4, the loss can not be converged to a satisfied results despite tuning various hyper-parameters.
Hi, I didn't try that in our experiments. I think this VQ-Stuff is unstable in terms of training, thus a 'proper' combination of these hyper-parameters is a must to make it work.
Hi, the default batch size is 1, did you try the larger batch size. In my experiments, when the batch size >= 4, the loss can not be converged to a satisfied results despite tuning various hyper-parameters.