Alert button
Picture for Dazhi Zhang

Dazhi Zhang

Alert button

SaaFormer: Spectral-spatial Axial Aggregation Transformer for Hyperspectral Image Classification

Jul 04, 2023
Enzhe Zhao, Zhichang Guo, Yao Li, Dazhi Zhang

Figure 1 for SaaFormer: Spectral-spatial Axial Aggregation Transformer for Hyperspectral Image Classification
Figure 2 for SaaFormer: Spectral-spatial Axial Aggregation Transformer for Hyperspectral Image Classification
Figure 3 for SaaFormer: Spectral-spatial Axial Aggregation Transformer for Hyperspectral Image Classification
Figure 4 for SaaFormer: Spectral-spatial Axial Aggregation Transformer for Hyperspectral Image Classification

Hyperspectral images (HSI) captured from earth observing satellites and aircraft is becoming increasingly important for applications in agriculture, environmental monitoring, mining, etc. Due to the limited available hyperspectral datasets, the pixel-wise random sampling is the most commonly used training-test dataset partition approach, which has significant overlap between samples in training and test datasets. Furthermore, our experimental observations indicates that regions with larger overlap often exhibit higher classification accuracy. Consequently, the pixel-wise random sampling approach poses a risk of data leakage. Thus, we propose a block-wise sampling method to minimize the potential for data leakage. Our experimental findings also confirm the presence of data leakage in models such as 2DCNN. Further, We propose a spectral-spatial axial aggregation transformer model, namely SaaFormer, to address the challenges associated with hyperspectral image classifier that considers HSI as long sequential three-dimensional images. The model comprises two primary components: axial aggregation attention and multi-level spectral-spatial extraction. The axial aggregation attention mechanism effectively exploits the continuity and correlation among spectral bands at each pixel position in hyperspectral images, while aggregating spatial dimension features. This enables SaaFormer to maintain high precision even under block-wise sampling. The multi-level spectral-spatial extraction structure is designed to capture the sensitivity of different material components to specific spectral bands, allowing the model to focus on a broader range of spectral details. The results on six publicly available datasets demonstrate that our model exhibits comparable performance when using random sampling, while significantly outperforming other methods when employing block-wise sampling partition.

* arXiv admin note: text overlap with arXiv:2107.02988 by other authors 
Viaarxiv icon

Stationary Point Losses for Robust Model

Feb 19, 2023
Weiwei Gao, Dazhi Zhang, Yao Li, Zhichang Guo, Ovanes Petrosian

Figure 1 for Stationary Point Losses for Robust Model
Figure 2 for Stationary Point Losses for Robust Model
Figure 3 for Stationary Point Losses for Robust Model
Figure 4 for Stationary Point Losses for Robust Model

The inability to guarantee robustness is one of the major obstacles to the application of deep learning models in security-demanding domains. We identify that the most commonly used cross-entropy (CE) loss does not guarantee robust boundary for neural networks. CE loss sharpens the neural network at the decision boundary to achieve a lower loss, rather than pushing the boundary to a more robust position. A robust boundary should be kept in the middle of samples from different classes, thus maximizing the margins from the boundary to the samples. We think this is due to the fact that CE loss has no stationary point. In this paper, we propose a family of new losses, called stationary point (SP) loss, which has at least one stationary point on the correct classification side. We proved that robust boundary can be guaranteed by SP loss without losing much accuracy. With SP loss, larger perturbations are required to generate adversarial examples. We demonstrate that robustness is improved under a variety of adversarial attacks by applying SP loss. Moreover, robust boundary learned by SP loss also performs well on imbalanced datasets.

* 13 pages, 12 figures 
Viaarxiv icon