Thanks for your impressive work.
During the inversion, I wondered why we need re-train a network to predict the camera pose (pitch & yaw).
Since there exist works focusing on predicting the head pose, can the predicted value from the other pre-trained models be used directly on the StyleNeRF?
However, in my experiment, I found that the predicted pose values range (DECA 3dmm) greatly differ from the camera pose used in StyleNeRF, which leads to poor inversion performance.
Thanks for your impressive work. During the inversion, I wondered why we need re-train a network to predict the camera pose (pitch & yaw). Since there exist works focusing on predicting the head pose, can the predicted value from the other pre-trained models be used directly on the StyleNeRF?
However, in my experiment, I found that the predicted pose values range (DECA 3dmm) greatly differ from the camera pose used in StyleNeRF, which leads to poor inversion performance.
Thanks in advance!