Thanks for the great work! I am currently delving into the functionality of the use_dynamic_rescale parameter in your project and have encountered a point of confusion that I hope you can clarify.
It appears that during the training phase, use_dynamic_rescale is applied to the input data xt (x = x * extract_into_tensor(self.scale_arr, t, x.shape)). However, during inference, it seems the rescaling is performed on the predicted x0(prev_scale_t = torch.full(size, self.ddim_scale_arr_prev[index], device=device)). This discrepancy—wherein the adjustment is made to the inputs during training but to the predictions at inference time—raises questions about the alignment between the training and inference processes. Is there any reference for this strategy?
Thanks for the great work! I am currently delving into the functionality of the use_dynamic_rescale parameter in your project and have encountered a point of confusion that I hope you can clarify.
It appears that during the training phase, use_dynamic_rescale is applied to the input data xt (x = x * extract_into_tensor(self.scale_arr, t, x.shape)). However, during inference, it seems the rescaling is performed on the predicted x0(prev_scale_t = torch.full(size, self.ddim_scale_arr_prev[index], device=device)). This discrepancy—wherein the adjustment is made to the inputs during training but to the predictions at inference time—raises questions about the alignment between the training and inference processes. Is there any reference for this strategy?