According to the paper, after passing the image through the Encoder, three feature maps of size 32x32 are obtained. After upsampling, the feature map size is 64x64, and PointFeatures are obtained through the Triplane Decoder. I would like to know what determines the number of PointFeatures input into the MLP and what the image size is after rendering.
According to the paper, after passing the image through the Encoder, three feature maps of size 32x32 are obtained. After upsampling, the feature map size is 64x64, and PointFeatures are obtained through the Triplane Decoder. I would like to know what determines the number of PointFeatures input into the MLP and what the image size is after rendering.