Closed lokvke closed 2 months ago
result.mp4
Do you use smooth_path?
@kike-0304 just train with the default config.
use smooth_path in test may have stable head, and would it be convenient for you to share your data processing code?
i tried to preprocess my own video data, the bs.npy is generated from EmoTalk project.
Hello, EmoTalk project does not provide data processing code. How did you generate the bs.np file?
i tried to preprocess my own video data, the bs.npy is generated from EmoTalk project.
Hello, EmoTalk project does not provide data processing code. How did you generate the bs.np file?
i don't know if it is right, maybe u can try it.
i tried to preprocess my own video data, the bs.npy is generated from EmoTalk project.
Hello, EmoTalk project does not provide data processing code. How did you generate the bs.np file?
i don't know if it is right, maybe u can try it.
This is bs.npy obtained from audio, we may need to obtain bs.npy from video frames
i tried to preprocess my own video data, the bs.npy is generated from EmoTalk project.
"face_rect": [ 61, 0, 384, 468 ] means [xmin, ymin, w, h]?
Hi @lokvke , may I ask which asr_model did you use to train your own video? if it's ave, how did you generate the aud_ave.npy file? thanks!
Hi @lokvke , may I ask which asr_model did you use to train your own video? if it's ave, how did you generate the aud_ave.npy file? thanks!
just use the provided audio_visual_encoder.pth
Thanks, I managed to train my own video, though the result is not as good as May's
Thanks, I managed to train my own video, though the result is not as good as May's
can u show your result here?
Thanks, I managed to train my own video, though the result is not as good as May's
can u show your result here?
bs.npy is closely related to the blinking action. Will your character blink when you replace the bs.npy? In your example video, the characters do not blink.
结果.mp4
Great, can you tell me how to get ave_npy?
I finally got it working:
https://github.com/ZiqiaoPeng/SyncTalk/assets/12045814/5e9e8a01-6c92-4827-b4aa-1bcb6db2499c
Thanks, I managed to train my own video, though the result is not as good as May's
can u show your result here?
bs.npy is closely related to the blinking action. Will your character blink when you replace the bs.npy? In your example video, the characters do not blink.
I updated code to only mask the lower part of the original video, so the characters' upper face remains the same as original video. See my sample above.
Thanks, I managed to train my own video, though the result is not as good as May's
can u show your result here?
bs.npy is closely related to the blinking action. Will your character blink when you replace the bs.npy? In your example video, the characters do not blink.
I updated code to only mask the lower part of the original video, so the characters' upper face remains the same as original video. See my sample above.
Could you share which part do you adjust for this? It looks awesome.
@jinqiupeter can you share the code adjustments you made for this? looks great.
hoping someone can create a short guide how to preprocess your own video data.
edit: welcome to join my discord server to discuss ai avatar stuff and sharing syncTalk tips https://discord.gg/jETUMmUD6h
@jinqiupeter can you share the code adjustments you made for this? looks great.
hoping someone can create a short guide how to preprocess your own video data.
edit: welcome to join my discord server to discuss ai avatar stuff and sharing syncTalk tips https://discord.gg/jETUMmUD6h
+1
@jinqiupeter can you share the code adjustments you made for this? looks great.
hoping someone can create a short guide how to preprocess your own video data.
edit: welcome to join my discord server to discuss ai avatar stuff and sharing syncTalk tips https://discord.gg/jETUMmUD6h
Preprocessing code is mostly the same as ER-NeRF
ER-NeRF
What audio feature extractor did you use?
ave, the code is already available in nerf_triplane/provider
i tried to preprocess my own video data, the bs.npy is generated from EmoTalk project.
Hello, EmoTalk project does not provide data processing code. How did you generate the bs.np file?
https://github.com/psyai-net/EmoTalk_release/blob/5179b27b2fdd1ca27fcbfa6a3264a5ecfd51d524/demo.py#L55 i don't know if it is right, maybe u can try it.
This is bs.npy obtained from audio, we may need to obtain bs.npy from video frames
hi, i use https://github.com/psyai-net/EmoTalk_release/blob/main/demo.py to create file bs.npy, but the shape output do not match with audio feature shape. Do you meet this problem, please give me advice , thanks.
Thanks, I managed to train my own video, though the result is not as good as May's
can u show your result here?
Hi can you share preprocessing code ? Thank for all your working !!!
result.mp4
@lokvke 大佬你好,请问如何训练自己的视频,你的预处理模块是否能提供下,我在自己的视频上进行训练,后续的优化效果模块与您持续共享,可以加我vx:wenfeng071555,期待与您的交流
I finally got it working:
kh_kr.mp4
Look great !! Can your share your preprocess and code
@jinqiupeter can you share the code adjustments you made for this? looks great. hoping someone can create a short guide how to preprocess your own video data. edit: welcome to join my discord server to discuss ai avatar stuff and sharing syncTalk tips https://discord.gg/jETUMmUD6h
Preprocessing code is mostly the same as ER-NeRF
Amazing result. What differences?
i tried to preprocess my own video data, the bs.npy is generated from EmoTalk project.