ZhexinLiang / CLIP-LIT

[ICCV 2023, Oral] Iterative Prompt Learning for Unsupervised Backlit Image Enhancement
https://zhexinliang.github.io/CLIP_LIT_page/
269 stars 23 forks source link

Questions about the training mechanism and loss design #18

Open JingyiXu404 opened 7 months ago

JingyiXu404 commented 7 months ago

Hi,

Very interesting work and I have some questions.

  1. Similarity of image features are constrainted by L_identity, I wonder why features are used instead of the back-lit and the enhanced image themselves? Is it because it will get better results?
  2. In your paper, the prompt pair and the enhancement part are trained alternatively. When studying the code, I was a little troubled. Could you please tell me how many iterations were trained for each step? Besides, I wonder why alternative training strategy is used instead of end-to-end train (For example, Loss = Lprompt+Lenhance)
  3. In prompt initialization, cross entropy loss is used with the designed labels, but in prompt finetuning process, the loss is changed. May you tell me why?

Thank you so much.

WangYuHzzz commented 5 months ago

I have similar questions, do you have an answer ,thank you