TY - JOUR
T1 - DA2Net
T2 - Distraction-Attention-Driven Adversarial Network for Robust Remote Sensing Image Scene Classification
AU - Yang, Rui
AU - Pu, Fangling
AU - Xu, Zhaozhuo
AU - DIng, Chujiang
AU - Xu, Xin
N1 - Publisher Copyright:
© 2004-2012 IEEE.
PY - 2022
Y1 - 2022
N2 - Optical remote sensing image (RSI) is easily affected by weather conditions. When the ground target is sheltered by clouds, extracting scene information from the RSI becomes quite challenging. In this work, we propose a distraction-attention-driven adversarial training network (DA2Net) to learn a robust RSI scene classification model. The distraction module employs a gradient-based class activation mapping (GradCAM++) method to produce partially occluded samples. Through feature map visualization, GradCAM++ can quantify the contribution of each region to the network prediction. Regions in the input image are erased and filled with white pixels if the corresponding contribution is higher than a given threshold. In this way, the distraction module enriches the training sample diversity and benefits the network's robustness and generalization performance. Training with the partially erased samples, the model can extract sufficient information from other regions even though the target with prominent features is occluded. The attention module highlights important features and information. It encourages the network to mine critical features from the uncovered regions. Competition between the two modules drives the network to improve its robustness and overall performance. Extensive experiments show that the DA2Net provides a promising approach for data augmentation and network training. Analysis of cloud-covered scene classification demonstrates the DA2Net's robust performance.
AB - Optical remote sensing image (RSI) is easily affected by weather conditions. When the ground target is sheltered by clouds, extracting scene information from the RSI becomes quite challenging. In this work, we propose a distraction-attention-driven adversarial training network (DA2Net) to learn a robust RSI scene classification model. The distraction module employs a gradient-based class activation mapping (GradCAM++) method to produce partially occluded samples. Through feature map visualization, GradCAM++ can quantify the contribution of each region to the network prediction. Regions in the input image are erased and filled with white pixels if the corresponding contribution is higher than a given threshold. In this way, the distraction module enriches the training sample diversity and benefits the network's robustness and generalization performance. Training with the partially erased samples, the model can extract sufficient information from other regions even though the target with prominent features is occluded. The attention module highlights important features and information. It encourages the network to mine critical features from the uncovered regions. Competition between the two modules drives the network to improve its robustness and overall performance. Extensive experiments show that the DA2Net provides a promising approach for data augmentation and network training. Analysis of cloud-covered scene classification demonstrates the DA2Net's robust performance.
KW - Adversarial training
KW - attention module
KW - distraction module
KW - remote sensing image (RSI) scene classification
UR - http://www.scopus.com/inward/record.url?scp=85107234974&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85107234974&partnerID=8YFLogxK
U2 - 10.1109/LGRS.2021.3079248
DO - 10.1109/LGRS.2021.3079248
M3 - Article
AN - SCOPUS:85107234974
SN - 1545-598X
VL - 19
JO - IEEE Geoscience and Remote Sensing Letters
JF - IEEE Geoscience and Remote Sensing Letters
ER -