Open 838959823 opened 1 year ago
We conducted extensive experiments on the location (before Enhance and Degrade modules) and number (whether Enhance module and Degrade module share the same attention map or not) of attention map IA, and found that the design we adopt in the paper performs best. We attribute it to two factors.
Our purpose is to enhance better, so we extract IA from low-light input IL to guide enhancement. As IA also guides degradation in the right, to obtain more realistic I~L, Mask Extractor (ME) is trained to be able to exploit more universal degradation features from IL.
Using specific IA for each Enhance and Degrade module needs more ME, making training more challenging to converge. You can see that as NeRCo has contained many components, we develop too many loss terms (i.e., Cooperative Loss (CL)) to constrain them. If we adopt more ME, the difficulty of training will inevitably increase.
I mean you can use the same ME. But given different image (unpaired like IL, IH), the attention map should be generated by the the same ME while using different input.
In this paper, ME is trained to depict low-light distribution of the degraded input. If inputting a clean image and degrade it following the guidance of ME, ME needs to be able to imagine possible dark regions based on normal-light images, which also increase training difficulty. At least our experiments show that this setting performs a little weaker than the seeting we final adopted.
i understand now. Thank you for your time!
In this picture, why IL and IH share the same attention map generated by Mask extractor when IL and IH are unpaired.