zhanglonghao1992 / One-Shot_Free-View_Neural_Talking_Head_Synthesis

Pytorch implementation of paper "One-Shot Free-View Neural Talking-Head Synthesis for Video Conferencing"
Other
782 stars 143 forks source link

Speeding up talking head synthesis #67

Closed yahskapar closed 1 year ago

yahskapar commented 1 year ago

I'm using a modified version of this project as a part of another project where talking head synthesis is a component. I'm curious if anyone here (author included) has any recommendations for speeding up the actual talking head synthesis itself, especially if it's used to augment, for example, tens of thousands of videos. I've achieved some moderate success just splitting up computation into different processes (given multiple processes running on GPUs with a larger memory pool, such as 48GB), but this is a bit clunky and also appears to mess up some videos by making them choppy at certain segments once stitched back together from multiple processes. Perhaps I missed it, but I don't really see any parameters to modify an inference-time batch size, or anything like that, within the code itself.

Any other ideas?

yahskapar commented 1 year ago

Closing this due to a lack of discussion.