Thank you for your great work!
I am trying to change camera poses at inference stage by inputting different camera_embedding to the pipeline - this gives me some unexpected results, so I tried to test a bit more about the camera_embedding parameter, and found some more unexpected results:
I firstly tried to input the default camera_embedding I found in mvdiffusion/pipelines/pipeline_mvdiffusion_image.py , and it works fine:
I wonder if this behaviour is expected? Does this mean that we are not supposed to change camera_embedding, and the output of inference step MUST be the 6 poses normal + the 6 poses rgb?
Thank you for your great work! I am trying to change camera poses at inference stage by inputting different
camera_embedding
to the pipeline - this gives me some unexpected results, so I tried to test a bit more about thecamera_embedding
parameter, and found some more unexpected results:I firstly tried to input the default
camera_embedding
I found in mvdiffusion/pipelines/pipeline_mvdiffusion_image.py , and it works fine:Code:
Result:
Then I tried to change the camera_embedding to the following, expecting to see the identical first image 12 times - but saw this different image:
Code:
Result:
And I tried some other combinations and they are worse:
Code:
Result:
Code:
Result:
I wonder if this behaviour is expected? Does this mean that we are not supposed to change
camera_embedding
, and the output of inference step MUST be the 6 poses normal + the 6 poses rgb?p.s., my full inference script is as follows:
Thank you so much! I look forward to your kind reply.