LizhenWangT / StyleAvatar

Code of SIGGRAPH 2023 Conference paper: StyleAvatar: Real-time Photo-realistic Portrait Avatar from a Single Video
BSD 2-Clause "Simplified" License
415 stars 45 forks source link

Input video instead of image for inference. #45

Open loic-combis opened 8 months ago

loic-combis commented 8 months ago

Hi there,

I'm trying to test style avatar to correct the lip sync generation from Wav2Lip.

(I currently have an issue with the preprocessing in Faceverse https://github.com/LizhenWangT/FaceVerse/issues/36)

But regardless, we aim to do the following:

Note that we try to avoid inputing a single image and recreate the whole speech on the avatar, and rather trying to simply correct the video.

Thanks for your help and your work!!