YuelangX / Gaussian-Head-Avatar

[CVPR 2024] Official repository for "Gaussian Head Avatar: Ultra High-fidelity Head Avatar via Dynamic Gaussians"
Other
776 stars 49 forks source link

dataset #22

Open Tiandishihua opened 6 months ago

Tiandishihua commented 6 months ago

Can I use RGB video from a monocular camera for training? We look forward to hearing from you, thank you.

YuelangX commented 6 months ago

Yes, you can follow the instruction to preprocess your monocular video for training. Actually, I tested on a set of monocular data. The rendering results from the side view look a lot worse.

https://github.com/YuelangX/Gaussian-Head-Avatar/assets/132422760/656912dc-98eb-4438-9e95-fc51eefa7a8d

https://github.com/YuelangX/Gaussian-Head-Avatar/assets/132422760/3b2182aa-d081-40eb-be2b-f2211a682f73

jeb0813 commented 6 months ago

Hi @YuelangX , for monocular data, how do you get the camera params(extrinsic, intrinsic).

YuelangX commented 6 months ago

You could refer to https://github.com/YuelangX/Multiview-3DMM-Fitting. The camera params are manually specified.

NikoBele1 commented 6 months ago

@YuelangX I am trying to use this repo with my monocular data not from Nersemble dataset.

Do you know a way to generate the camera params for the dataset structure required in your repo Multiview-3DMM-Fitting ? As in, frame1.jpg - camera1.npz, frame2.jpg - camera2.npz, etc.

YuelangX commented 6 months ago

@NikoBele1 , you can run this script https://github.com/YuelangX/Multiview-3DMM-Fitting/blob/6f64ee1bb1545ce3f3a83706f07f857b0f28ab5d/preprocess/preprocess_monocular_video.py

NikoBele1 commented 6 months ago

oh my bad, i thought the camera params where required, i misread that the repository creates them.

NikoBele1 commented 6 months ago

@YuelangX sorry to disturb you again, but i have a question. I am currently training on a monocular video (539 frames, preprocessed with Multiview-3DMM-Fitting) and am at 600 epochs. I used checkpoint 600 to do self-reenactment on my monocular video and this is the result i get:

intermediate_GHA_1_720

This does not look like Gaussian Splatting, i would expect to see larger splats everywhere. Problems i know: The camera params are not correct, since they are manual in Multiview-3DMM-Fitting/preprocess_monocular_video.py. And i noticed that there is no lowres landmark in myDataset/mySubject/landmarks/*/. after preprocess.

Do you have an idea what could be wrong here? I know training is not finished, but i would expect to see a different intermediate result.

YuelangX commented 6 months ago

@NikoBele1 This seems strange, are the results during training also like this?

NikoBele1 commented 5 months ago

@YuelangX thanks for replying. Tracking looks fine, besides the strange cropping. This is probably because of the manual camera parameters i used from Multiview-3DMM-Fitting/preprocess_monocular_video.py ? 413900

jryebread commented 4 months ago

@NikoBele1 hi did you figure out the fix to your issue for custom character? have same issue

NikoBele1 commented 4 months ago

@jryebread what issue do you have? the cropping or the other weird reenactment?