HQ-50K: A Large-scale, High-quality Dataset for Image Restoration. The repository is for our paper HQ-50K: A Large-scale, High-quality Dataset for Image Restoration.
Paper | Dataset | [Pretrained models]()
50,000 high-quality images with rich texture details and semantic diversity
HQ-50K a large-scale and high-quality image restoration dataset which contains 50,000 high-quality images with rich texture details and semantic diversity, considering the five aspects simultaneously : Large-Scale, High-Resolution, Compression Rates, Rich texture details and Semantic Coverage. We also present a new Degradation-Aware Mixture of Expert (DAMoE) model, which enables a single model to handle multiple corruption types and unknown levels.
In addition to the 50,000 images for training, we also offer 1250 test images that span across various semantic categories and frequency ranges. This new benchmark can facilitate detailed and fine-grained performance comparison and analysis.
│HQ-50K/
├──train/
│ ├── the first image url
│ ├── the second image url
│ ├── ......
│ ├── ......
│ ├── 50000th image url
├──val/
│ ├── animal
│ │ ├── ......
│ ├── architcture
│ │ ├── ......
│ ├── ......
If the full dataset is hard to available due to packet loss, we provide [alternative ways](https://github.com/littleYaang/HQ-50K/blob/main/data_preprocess/README.md) for dataset download.
2. Generate paired data corresponding to each task [DataPrepare]().
## DAMoE Model
<img src="https://github.com/littleYaang/HQ-50K/raw/main/figures/backbone.png" width="800px"/>
Coming Soon
## Cite HQ-50K
@misc{yang2023hq50k, title={HQ-50K: A Large-scale, High-quality Dataset for Image Restoration}, author={Qinhong Yang and Dongdong Chen and Zhentao Tan and Qiankun Liu and Qi Chu and Jianmin Bao and Lu Yuan and Gang Hua and Nenghai Yu}, year={2023}, eprint={2306.05390}, archivePrefix={arXiv}, primaryClass={cs.CV} }
## License and Acknowledgement
The dataset is released under the CC-BY-4.0 license. And the codes are based on KAIR and Fastmoe. Please also follow their licenses. Thanks for their awesome works.