Open AlexCHENSiyu opened 5 months ago
ok, probably I find the reason. BTW , this part should be written like this: add ['arr_0'] to it. otherwise cannot run.
I had this when I did not clamp styled image: https://github.com/ivanstepanovftw/artistic-style-transfer-for-videos/blob/b42b79d17fd1b72917b55100182c6f5f3d097d92/main_video_join.py#L57 Similar as it is performed in the closure: https://github.com/ivanstepanovftw/artistic-style-transfer-for-videos/blob/b42b79d17fd1b72917b55100182c6f5f3d097d92/main_video.py#L342 I actually thinking about adding loss for out of bounds pixels, so the optimizer would not be focused on these pixels.
yes, right after I posted the issue, I found this mistake of mine. I read the paper carefully and I realize that long-term temporal consistency requires heavy-duty computation. It might be hard for me to achieve that.
BTW, do you know any comparative standard to measure the quality of style-transferred images between different pre-train model?
and any reason why you choose efficientnet-b0 as the best model
BTW, do you know any comparative standard to measure the quality of style-transferred images between different pre-train model?
Unfortunately, no.
and any reason why you choose efficientnet-b0 as the best model
I am not sure if it is best for style transfer. EfficientNet chosen as it is efficient, since I do not have local GPU.
Also, in the paper "CNN Filter DB: An Empirical Investigation of Trained Convolutional Filters" (arxiv:2203.15331v2), they say that "learned filters do not significantly differ across models trained for various tasks, except for extreme outliers such as GAN-Discriminators."
I can get this good output yesterday. but today when I ran it again, I get this strange output:
Have you ever encountered this problem before? Or you can give me some suggestions?