학술논문

Spatial-Gated Multilayer Perceptron for Land Use and Land Cover Mapping
Document Type
Periodical
Source
IEEE Geoscience and Remote Sensing Letters IEEE Geosci. Remote Sensing Lett. Geoscience and Remote Sensing Letters, IEEE. 21:1-5 2024
Subject
Geoscience
Power, Energy and Industry Applications
Communication, Networking and Broadcast Technologies
Components, Circuits, Devices and Systems
Computing and Processing
Signal Processing and Analysis
Feature extraction
Classification algorithms
Hyperspectral imaging
Data models
Transformers
Biological system modeling
Training data
Attention mechanism
image classification
spatial gating unit (SGU)
vision transformers (ViTs)
Language
ISSN
1545-598X
1558-0571
Abstract
Due to its capacity to recognize detailed spectral differences, hyperspectral (HS) data have been extensively used for precise land use land cover (LULC) mapping. However, recent multimodal methods have shown their superior classification performance over the algorithms that use single datasets. On the other hand, convolutional neural networks (CNNs) are models extensively utilized for the hierarchical extraction of features. Vision transformers (ViTs), through a self-attention mechanism, have recently achieved superior modeling of global contextual information compared to CNNs. However, to harness their image classification strength, ViTs require substantial training datasets. In cases where the available training data is limited, current advanced multilayer perceptrons (MLPs) can provide viable alternatives to both deep CNNs and ViTs. In this letter, we developed the SGU-MLP, a deep-learning algorithm that effectively combines MLPs and spatial gating units (SGUs) for precise LULC mapping using multimodal data from multispectral, LiDAR, and HS data. Results illustrated the superiority of the developed SGU-MLP classification algorithm over several CNN- and CNN-ViT-based models, including HybridSN, ResNet, iFormer, EfficientFormer, and CoAtNet. The SGU-MLP classification model consistently outperformed the benchmark CNN- and CNN-ViT-based algorithms. The code will be made publicly available at https://github.com/aj1365/SGUMLP.