Closed iperov closed 3 years ago
Hi, thanks for your question. As stated in the Readme, we have slightly modified the training process and get much higher performance. Adding the original high-freq can benefit the initialization of D by avoiding discriminating totally broken images without introducing parameters. The overall pipeline is unchanged where a progressive masking strategy saves much computation on high-freq yet combines deep information from the learning of masks. Removing the mentioned residual operation can get 22.7db at 480p.
- First Difference: The paper use conv-lrelu-conv refine the mask, but the code don't refine the mask. so the three high branch can parallel compute. This here diff so much.
- Second Difference: as @iperov said;
- channel information is different, the channel of mask in paper is 1. did it a write error?
Hi, sorry for confusing you. We have uploaded the LPTN_paper_arch.py
and train_FiveK_paper.yml
that directly implements the model illustrated in our paper. This model achieves 22.3db at 480p in this code. For your questions:
cannot understand this block
here you multiply mask from first high freq model with original high freq map
but then
add
original high freq map?I don't see that in the paper