A large and diverse HSI dataset named HSIHybrid was curated for large-scale HSI pre-training. It consisted of 15 HSI datasets from different hyperspectral sensors. After splitting into image patches, a total of 4 million HSI patches with a spatial size of 9×9 were obtained.
A group-wise PCA was used to extract features of HSI spectra and transform the raw spectra to fixed-length features.
A modified MAE named HSIMAE that utilized separate spatial-spectral encoders followed by fusion blocks to learn spatial correlation and spectral correlation of HSI data was proposed.
A dual-branch fine-tuning framework was introduced to leverage the unlabeled data of the downstream HSI dataset and suppressed overfitting on small training samples.
Install Git
Open commond line, create environment and enter with the following commands:
conda create -n HSIMAE python=3.8
conda activate HSIMAE
Clone the repository and enter:
git clone https://github.com/Ryan21wy/HSIMAE.git
cd HSIMAE
Install dependency with the following commands:
pip install -r requirements.txt
The pre-training dataset and pretrained models of HSIMAE are provided in Hugging Face.
Because it is too big, HySpecNet-11k need be downloaded from HySpecNet-11k - A Large-Scale Hyperspectral Benchmark Dataset (rsim.berlin)
Salinas: Salinas scene
Pavia University: Pavia University
Houston 2013: 2013 IEEE GRSS Data Fusion Contest
WHU-Hi-LongKou: WHU-Hi: UAV-borne hyperspectral and high spatial resolution (H2) benchmark datasets
Overall accuracy of four HSI classification datasets. The training set and validation set contained 5/10/15/20 random samples per class , respectively, and the remaining samples were considered as the test set.
Training Samples | Salinas | Pavia University | Houston 2013 | WHU-Hi-LongKou | Average |
---|---|---|---|---|---|
5 | 92.99 | 87.00 | 83.89 | 96.16 | 90.01 |
10 | 95.14 | 96.02 | 90.14 | 97.64 | 94.74 |
15 | 96.51 | 97.09 | 94.52 | 98.08 | 96.55 |
20 | 96.62 | 97.44 | 95.65 | 98.41 | 97.03 |
If you think this project is helpful, please feel free to leave a star⭐️ and cite our paper:
@ARTICLE{10607879,
author={Wang, Yue and Wen, Ming and Zhang, Hailiang and Sun, Jinyu and Yang, Qiong and Zhang, Zhimin and Lu, Hongmei},
journal={IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing},
title={HSIMAE: A Unified Masked Autoencoder with Large-scale Pre-training for Hyperspectral Image Classification},
year={2024},
doi={10.1109/JSTARS.2024.3432743}
}
Wang Yue
E-mail: ryanwy@csu.edu.cn