oppo-us-research / SpacetimeGaussians

[CVPR 2024] Spacetime Gaussian Feature Splatting for Real-Time Dynamic View Synthesis
https://oppo-us-research.github.io/SpacetimeGaussians-website/
Other
616 stars 45 forks source link

Custom data / Technicolor dataset structure example #16

Closed henrypearce4D closed 9 months ago

henrypearce4D commented 9 months ago

Hi great work with this project!

I have request access to the Light-filed dataset. In case I am not given access, would it be possible to provide a breakdown of the folder structure and the files included? It looks like they use image sequences rather than .mp4, this is what we would like to test with spacetime gaussians.

henrypearce4D commented 9 months ago

I was given access, the undistorted dataset contains;

cameras_parameters.txt
# f cu cv ar sk  qw qx qy qz  tx ty tz

Image convention;

Fabien_undist_[Frame Number]_[Camera Name].png
e.g.
Fabien_undist_00001_00.png
Fabien_undist_00001_01.png
Fabien_undist_00001_02.png
Fabien_undist_00002_00.png
Fabien_undist_00002_01.png
Fabien_undist_00002_02.png
etc
aleatorydialogue commented 7 months ago

Were you able to get any custom video to work with these conventions? I'm trying to find a way to train with monocular video all day with no luck.

henrypearce4D commented 7 months ago

@aleatorydialogue I might be mistaken but this work is not intended with monocular video in mind. All the examples are multi-camera.

aleatorydialogue commented 7 months ago

I think you are right. I had hoped I could adapt it to my own data, even if lower quality result. I'm mostly interested in usijg splatV viewer, which only seems to work with output from this project. I am trying to work on using 4dgs loader to get it to train with just one camera, but no success yet.

lizhan17 commented 7 months ago

I think you are right. I had hoped I could adapt it to my own data, even if lower quality result. I'm mostly interested in usijg splatV viewer, which only seems to work with output from this project. I am trying to work on using 4dgs loader to get it to train with just one camera, but no success yet.

although our method is not designed for one camera. But it can converge with the "mono" camera used in 4dgs or deformable 3dgs. As the dataset used in these two papers are one camera that "moving" around objects. Just use the initial points used in 4dgs and replicate points into different time stamps, which can be seen as multi camera captured at different times.
You can still get similar results as these two papers with some minor modifications (this is a future work).