Shuai Yang$^{{*}}$, [ZhiFei Chen]()$^{{*}}$, [Pengguang Chen](), [Xi Fang](), [Yixun Liang](), [Shu Liu](), Yingcong Chen$^{**}$
HKUST(GZ), HKUST, SmartMore Corp.
${*}$: Equal contribution. **: Corresponding author.
Industrial datasets often lack detailed defect annotations, providing only binary masks or misclassifications. We introduce the Defect Spectrum, a comprehensive dataset with refined, large-scale annotations for various industrial defects. Using four industrial benchmarks, Defect Spectrum enhances annotation accuracy, capturing subtle and previously missed defects. Our dataset includes rich semantic annotations, identifying multiple defect types per image, and offers descriptive captions for each sample, facilitating future Vision Language Model research.
Furthermore, we introduce Defect-Gen, a two-stage diffusion-based generator designed to create high-quality and diverse defective images, even when working with limited defective data. The synthetic images generated by Defect-Gen significantly enhance the performance of defect segmentation models, achieving an improvement in mIoU scores up to 9.85 on Defect-Spectrum subsets.
This generative model excels in producing diverse and high-quality images, even when trained on limited data.
conda create -n diff python==3.8.0
.conda activate diff
.pip install -r requirements.txt
.train_[large/small].sh
that corresponds to your own needs. e.g. If the "Capsule" object has 7 defective classes, set the --num_defect
to 7. -input channel
and -output channel
should be a total of the number of defect types, RGB channels, and background channels(if needed). e.g. For an object that has 7 defective classes, the number of input/output channels should be set to 10. (excluding background)sh train_[large/small].sh
/[working_dir]/checkpoint
.inference.sh
.--step_inference
. --num_defect
.sh inference.sh
.If you find this project useful in your research, please consider citing:
@misc{yang2023defect,
title={Defect Spectrum: A Granular Look of Large-Scale Defect Datasets with Rich Semantics},
author={Shuai Yang and Zhifei Chen and Pengguang Chen and Xi Fang and Shu Liu and Yingcong Chen},
year={2023},
eprint={2310.17316},
archivePrefix={arXiv},
primaryClass={cs.CV}
}