Closed kimwao closed 1 year ago
About pretrained models,
cunet/
and upconv_7/
are converted from the original waifu2x pretrained model to pytorch. In other words, they are trained with the original waifu2x code and the dataset at the time of their release. The reason is for compatibility.
Only swin_unet/
is trained with this repository and the latest dataset.
I have also tried to train upconv_7
and upcunet
with this repository and the latest dataset.
(These are not included in the release. Also, I have changed the code even in the last few days so it is not exactly the same.)
Benchmark result
python3 -m waifu2x.benchmark --model-dir ${MODEL_DIR} --method scale -i /data/dataset/eval --filter catrom --color y_matlab
(No TTA)
arch | trained with nunif | trained with the original waifu2x |
---|---|---|
upconv_7 | 37.6632 | 37.0478 |
upcunet | 39.9458 | 39.458 |
swin_unet_2x | 40.3235 | N/A |
I don't know if someone else already asked this or if I just got lost in the documentation, but it seems that it is not easy to find the new database and the differences between the original and the updated one, is there a quick link to verify this?
I use hybrid of self crawled dataset and synthetic dataset(font and dot patterns), which I change from time to time on a whim, and I have never described it in detail.
upconv_7, upcunet, swin_unet_2x are .pth? or can they be ported? do they work like unlimited:waifu2x? waifu2x.udp? waifu2x could be the best upscaler for anime, better than esgran anime-6b, animevideo or v3, waifuxl did this claim 1 year ago : Real-ESRGAN will outperform the models used on [waifu2x], but i think its not true, your upscalers with Deep Convolutional Neural Networks have better potential, upscaling and upscalesr are becoming the most important part of artificial generated images, like swinir, esgran,scunet,hat, you can load the .pth to upscale encode/decode/vae , pkuliyi2015, multidiffusion extension is a good example, you can create an extension for tools like automatic1111, or release arch checkpoints with nuninf/waifu2x fo testing, yu45020/Waifu2x checkpointsare outdated, antonpaquin/waifu2x-cunet-pytorch) outdated too, and emulations waifu2x.pth dont work well, i really think this upscaler is the future.
@ladichanjp The closest thing to a compendium of anime upscaling is vs-mlrt, where they have from the traditional waifu2x cunet and upconv 7 and Real ESRGAN, to the recent waifu2x swin unet and Real CUGAN. The way they are compiled is through ports to ONNX, which has allowed them to use traditional CUDA acceleration, to take advantage of vulkan or rt cores. The big disadvantage is that you must at least know a little python and vapoursynth, which are the basis of this whole operation.
@ladichanjp The closest thing to a compendium of anime upscaling is vs-mlrt, where they have from the traditional waifu2x cunet and upconv 7 and Real ESRGAN, to the recent waifu2x swin unet and Real CUGAN. The way they are compiled is through ports to ONNX, which has allowed them to use traditional CUDA acceleration, to take advantage of vulkan or rt cores. The big disadvantage is that you must at least know a little python and vapoursynth, which are the basis of this whole operation.
thank you so much, this is very intersting, automatic1111 Onnx Pipeline is not ready, is there a way to port that to a pth format? VapourSynth looks amazing, its going to be essential to upscale txt2video
@ladichanjp I realized I forgot that there was already a GUI front end for this whole vs mlrt if you are interested, enhancr About going to the original pth, I think you might have to dig into every project. Regarding waifu2x, nagadomi already offers pre trained models which are already on the pth format you are looking for. But, unless you are already onto python and messing with ai libraries, I think you should be safe with enhancr.
@ladichanjp I realized I forgot that there was already a GUI front end for this whole vs mlrt if you are interested, enhancr About going to the original pth, I think you might have to dig into every project. Regarding waifu2x, nagadomi already offers pre trained models which are already on the pth format you are looking for. But, unless you are already onto python and messing with ai libraries, I think you should be safe with enhancr.
thanks you so much xurich-xulaco and nagadomi, this will help me with my research
If you are using Stable Diffusion Web UI, you should be able to run Command Line Interface or Web UI of this repo. If you want it to work in Colab, see #9 .
Also, I am recently working on a model for photos. It is also designed for photorealistic generated art like Stable diffusion, so if it works out, I may develop an extension for sd. However, that does not mean right now, so if someone wants to make it, you can make it.
Thanks for your excellent work! I want to verify if the increase in the network's parameters can improve performance, so I need to determine if they were trained from the same dataset. I would be very grateful if you tell me!