Dichao-Liu / CMAL

44 stars 11 forks source link

CMAL-Net

Code release for 'Learn from Each Other to Classify Better: Cross-layer Mutual Attention Learning for Fine-grained Visual Classification'. You may check more details in our paper published in Pattern Recognition if you are interested in our work.

enter image description here

Environment

This source code was tested in the following environment:

Python = 3.7.11

PyTorch = 1.8.0

torchvision = 0.9.0

Ubuntu 18.04.5 LTS

NVIDIA GeForce RTX 3080 Ti

Dataset

Dependencies

Install Inplace-ABN following the instructions:

https://github.com/Alibaba-MIIL/TResNet/blob/master/requirements.txt

https://github.com/Alibaba-MIIL/TResNet/blob/master/INPLACE_ABN_TIPS.md

Download the folder src from https://github.com/Alibaba-MIIL/TResNet and save it as:

Code
├── basic_conv.py
├── utils.py
├── train_Stanford_Cars_TResNet_L.py
├── ...
├── src

Training

Run the scripts for training, such as python train_Stanford_Cars_TResNet_L.py.

Acknowledgement

Part of the training code is inspired by Du et al, and many thanks to them.

Citation

@article{LIU2023109550,
title = {Learn from each other to Classify better: Cross-layer mutual attention learning for fine-grained visual classification},
journal = {Pattern Recognition},
volume = {140},
pages = {109550},
year = {2023},
issn = {0031-3203},
doi = {https://doi.org/10.1016/j.patcog.2023.109550},
url = {https://www.sciencedirect.com/science/article/pii/S0031320323002509},
author = {Dichao Liu and Longjiao Zhao and Yu Wang and Jien Kato}
}