G-U-N / AnimateLCM

AnimateLCM: Let's Accelerate the Video Generation within 4 Steps!
https://animatelcm.github.io
MIT License
567 stars 41 forks source link

Teacher-Free Adaptation is Latent Consistency Fine-tuning (LCF)? #10

Closed dcfucheng closed 6 months ago

dcfucheng commented 6 months ago

Great works~

I am confused about 'Teacher-Free Adaptation'. Does it mean Latent Consistency Fine-tuning (LCF)? Directly select two time steps and get nosied data z{t} and z{t-1}, and then directly calculate the consistency loss for these two time steps to enforce self-consistency property as LCF in LCM paper?

So the training procedure is :

  1. Train base image diffusion model, using Latent Consistency Distillation and image data.
  2. Fix the lcm image diffusion weight and add a trainable temporal layer, using Latent Consistency Distillation and new Initialization strategy and video data.
  3. Add other controls (controlnet or IPadpter), using Latent Consistency Fine-tuning and the data about control conditions.

Do I have an exact understanding?

G-U-N commented 6 months ago

Good summary! Thanks for your interest and careful reading. For the Teacher-Free Adaptation and Latent Consistency Fine-tuning, I would say that they are technically the same thing, all following the consistency-training (CT) proposed in consistency models by Song et al. The difference is that we verify that it works for training adapters from scratch or fine-tuning existing adapters to follow consistency property, mainly at the purpose level.