Closed A-2-H closed 1 month ago
It seems the problem is in EchoMimic\src\pipelines\pipeline_echo_mimic_pose_acc.py
@torch.no_grad()
def __call__(
..................
print("23: with self.progress_bar")
with self.progress_bar(total=num_inference_steps) as progress_bar:
print(f"23.1: total:{total}, num_inference_steps:{num_inference_steps}")
Output
(echomimic) C:\tut\EchoMimic>python -u infer_audio2vid_pose_acc.py
C:\Users\nitin\miniconda3\envs\echomimic\lib\site-packages\diffusers\utils\outputs.py:63: UserWarning: torch.utils._pytree._register_pytree_node is deprecated. Please use torch.utils._pytree.register_pytree_node instead.
torch.utils._pytree._register_pytree_node(
C:\Users\nitin\miniconda3\envs\echomimic\lib\site-packages\diffusers\utils\outputs.py:63: UserWarning: torch.utils._pytree._register_pytree_node is deprecated. Please use torch.utils._pytree.register_pytree_node instead.
torch.utils._pytree._register_pytree_node(
23: with self.progress_bar
(echomimic) C:\tut\EchoMimic>
It is not going inside with statement.
If developers can help how to debug I am ready to help.
Some more logs. Not going inside with statement
(echomimic) C:\tut\EchoMimic>python -u infer_audio2vid_pose_acc.py
C:\Users\nitin\miniconda3\envs\echomimic\lib\site-packages\diffusers\utils\outputs.py:63: UserWarning: torch.utils._pytree._register_pytree_node is deprecated. Please use torch.utils._pytree.register_pytree_node instead.
torch.utils._pytree._register_pytree_node(
C:\Users\nitin\miniconda3\envs\echomimic\lib\site-packages\diffusers\utils\outputs.py:63: UserWarning: torch.utils._pytree._register_pytree_node is deprecated. Please use torch.utils._pytree.register_pytree_node instead.
torch.utils._pytree._register_pytree_node(
1: Initializing pipeline
2: Pipeline initialized
7: Getting execution device
video in 24 FPS, audio idx in 50FPS
15: Preparing latents
17: latents shape:torch.Size([1, 4, 160, 64, 64]), video_length:160
20: face_locator_tensor
10: Preparing extra step kwargs
11: Extra step kwargs prepared
21: extra_step_kwargs
22: denoising loop
23: with self.progress_bar
ref_image_latents shape: torch.Size([1, 4, 64, 64])
face_mask_tensor shape: torch.Size([1, 3, 240, 512, 512])
face_locator_tensor shape: torch.Size([1, 320, 240, 64, 64])
self.progress_bar: <bound method DiffusionPipeline.progress_bar of AudioPose2VideoPipeline {
"_class_name": "AudioPose2VideoPipeline",
"_diffusers_version": "0.24.0",
"audio_guider": [
"src.models.whisper.audio2feature",
"Audio2Feature"
],
"denoising_unet": [
"src.models.unet_3d_echo",
"EchoUNet3DConditionModel"
],
"face_locator": [
"src.models.face_locator",
"FaceLocator"
],
"image_proj_model": [
null,
null
],
"reference_unet": [
"src.models.unet_2d_condition",
"UNet2DConditionModel"
],
"scheduler": [
"diffusers",
"DDIMScheduler"
],
"text_encoder": [
null,
null
],
"tokenizer": [
null,
null
],
"vae": [
"diffusers",
"AutoencoderKL"
]
}
>
(echomimic) C:\tut\EchoMimic>
Code
print("23: with self.progress_bar")
print("ref_image_latents shape:", ref_image_latents.shape)
print("face_mask_tensor shape:", face_mask_tensor.shape)
print("face_locator_tensor shape:", face_locator_tensor.shape)
print("self.progress_bar:", self.progress_bar)
with self.progress_bar(total=num_inference_steps) as progress_bar:
print("Inside with statement")
maybe Motion Sync just for help people extract motion pkl file from video? you'll get new dir to save pkl files with run this script. and you can run audio2vid_pose to sync it well , right?
maybe Motion Sync just for help people extract motion pkl file from video? you'll get new dir to save pkl files with run this script. and you can run audio2vid_pose to sync it well , right?
We are just trying to inference sample provided in this repo which is not working.
Creating pickle using our own video is working fine.
motion sync only provides pkl files for each frame. It is a pre-process for the driven video (if you have your own driven video and ref image, you should run it before calling infer to generate video). Now, try the new script released. Motion sync is done online in the infer process. No need to run it individually.
Windows 10, conda. I tried audio2video script before and it worked. Now that I updated your program to try "motion sync" nothing happen. I tried multiple times the motion sync script:
in my root folder EchoMimic created folder named "d" it's the same name as source image. I can see some ".pkl" files but that's it. Overall there is 288 pkl files in "d" folder. So it matches the numer in conda "288". But still it didn't rendered video.
This is my pip list: