Open Symbolzzz opened 3 weeks ago
Hello,
Please refer to here. https://github.com/ai4r/Gesture-Generation-from-Trimodal-Context/blob/master/scripts/synthesize.py#L121 It demonstrates motion synthesis for a long speech.
Thank you for your reply. It helps me a lot.🙏
Another question, I referred to the smooth transition method in the code, and I found that the synthesized short sequence fragments still appear incoherent. Is there a good solution for this?
It was a faily simple transtion smoothing, and there is room for improvement. I don't have a good and easy solution, but you can try adjusting the hyperparameters of generation length or number of seed poses for better transition.
Thanks again for your patient reply!
Very good framework.
I noticed the description of seed pose in the article:
But I found in the code that it seems that only the first four frames of the current GT are used as seed poses.
I want to know how to splice the predicted long sequence into a complete long sequence after slicing them into short sequences of 34 frames for a long input sequence. I am currently referring to your code framework in my work, but I encountered the problem of incoherent generated gestures when splicing the predicted short sequence of 34 frames. Is this related to the seed pose or the splicing method?
I would be grateful if you could answer my question.