Code of SIGGRAPH 2023 Conference paper: StyleAvatar: Real-time Photo-realistic Portrait Avatar from a Single Video
BSD 2-Clause "Simplified" License
415
stars
45
forks
source link
Input video instead of image for inference. #45
Open
loic-combis opened 8 months ago
Hi there,
I'm trying to test style avatar to correct the lip sync generation from Wav2Lip.
(I currently have an issue with the preprocessing in Faceverse https://github.com/LizhenWangT/FaceVerse/issues/36)
But regardless, we aim to do the following:
Note that we try to avoid inputing a single image and recreate the whole speech on the avatar, and rather trying to simply correct the video.
Thanks for your help and your work!!