Open guibo-tech opened 1 year ago
GPU runs out of memory, using CPU solve it,
1st evaluation using the new generated images, all gray, why? reco_0003_training_2
python3 EndoNeRF/eval_rgb.py --gt_dir /home/guibo/repos/reco_0003/images --mask_dir /home/guibo/repos/reco_0003/gt_masks --img_dir logs/reco_0003_training_2/renderonly_path_fixidentity_000000/estim/
MSE: 0.05545433238148689 PSNR: 12.56064510345459 SSIM: 0.6856867074966431 LPIPS: 0.50532066822052
reco_0003_training_2
python3 EndoNeRF/eval_rgb.py --gt_dir /home/guibo/repos/reco_0003/images --mask_dir /home/guibo/repos/reco_0003/gt_masks --img_dir /home/guibo/repos/BIMAP-EndoNeRF/EndoNeRF/logs/reco_0003_training_2/frames_reco_0003_training_2_fixidentity_100000_time/estim
MSE: 0.0005877256044186652 PSNR: 32.30825424194336 SSIM: 0.889693021774292 LPIPS: 0.276197612285614
reco_0003_training_1
python3 EndoNeRF/eval_rgb.py --gt_dir /home/guibo/repos/reco_0003/images --mask_dir /home/guibo/repos/reco_0003/gt_masks --img_dir /home/guibo/repos/frames_reco_0003_training_1_fixidentity_100000_time/estim
MSE: 0.0005817899946123362 PSNR: 32.35233688354492 SSIM: 0.8909807801246643 LPIPS: 0.27124789357185364
Comparing the two reconstructions.
In both cases, the differences in the metrics between Training 1 and Training 2 are relatively small. Prioritizing traditional image quality metrics like PSNR and SSIM, Training 1 has a slight advantage. On the other hand, prioritizing perceptual similarity, Training 1 has a slightly lower LPIPS.
MSE (Mean Squared Error):
PSNR (Peak Signal-to-Noise Ratio):
SSIM (Structural Similarity Index):
LPIPS (Learned Perceptual Image Patch Similarity):
The updated results you've provided are as follows:
These results indicate the quality and similarity of the estimated images compared to the ground truth images. Here's a brief summary of each metric:
MSE (Mean Squared Error): MSE measures the average squared difference between the estimated and ground truth images. A lower MSE indicates higher image quality. An MSE of 0.055 suggests relatively low error or distortion in the estimated images.
PSNR (Peak Signal-to-Noise Ratio): PSNR measures image quality based on the ratio of the peak signal (maximum possible value) to noise. A higher PSNR indicates higher image quality. A PSNR of 12.56 suggests that there is still room for improvement in image quality.
SSIM (Structural Similarity Index): SSIM measures the structural similarity between two images. A higher SSIM value indicates that the two images are more similar in terms of structure and content. An SSIM of 0.686 suggests that the estimated images are somewhat similar to the ground truth.
LPIPS (Learned Perceptual Image Patch Similarity): LPIPS is a perceptual similarity metric. A lower LPIPS value indicates higher similarity. An LPIPS of 0.505 suggests that there is room for improvement in terms of perceptual similarity between the estimated and ground truth images.
These metrics provide quantitative insights into the quality and similarity of the images, but the interpretation of these results depends on the specific application and the desired level of image quality. Depending on your use case, you may want to further fine-tune your model to improve these metrics or assess the visual quality of the estimated images.
Error incompatibility with images and gt_images If there are 45 images and 45 gt_images, we get the following error. Running the code 45 images and 43 gt_images are loaded. To solve it, just copy two images more in the gt_images folder.
The same error we see to train the EndoNeRF, we have to copy two images more in the folder masks