[go: up one dir, main page]

 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,265)

Search Parameters:
Keywords = hyperspectral image classification

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
17 pages, 5425 KiB  
Article
HyperKon: A Self-Supervised Contrastive Network for Hyperspectral Image Analysis
by Daniel La’ah Ayuba, Jean-Yves Guillemaut, Belen Marti-Cardona and Oscar Mendez
Remote Sens. 2024, 16(18), 3399; https://doi.org/10.3390/rs16183399 - 12 Sep 2024
Abstract
The use of a pretrained image classification model (trained on cats and dogs, for example) as a perceptual loss function for hyperspectral super-resolution and pansharpening tasks is surprisingly effective. However, RGB-based networks do not take full advantage of the spectral information in hyperspectral [...] Read more.
The use of a pretrained image classification model (trained on cats and dogs, for example) as a perceptual loss function for hyperspectral super-resolution and pansharpening tasks is surprisingly effective. However, RGB-based networks do not take full advantage of the spectral information in hyperspectral data. This inspired the creation of HyperKon, a dedicated hyperspectral Convolutional Neural Network backbone built with self-supervised contrastive representation learning. HyperKon uniquely leverages the high spectral continuity, range, and resolution of hyperspectral data through a spectral attention mechanism. We also perform a thorough ablation study on different kinds of layers, showing their performance in understanding hyperspectral layers. Notably, HyperKon achieves a remarkable 98% Top-1 retrieval accuracy and surpasses traditional RGB-trained backbones in both pansharpening and image classification tasks. These results highlight the potential of hyperspectral-native backbones and herald a paradigm shift in hyperspectral image analysis. Full article
(This article belongs to the Special Issue Advances in Hyperspectral Remote Sensing Image Processing)
18 pages, 10993 KiB  
Article
Hyperspectral Rock Classification Method Based on Spatial-Spectral Multidimensional Feature Fusion
by Shixian Cao, Wenyuan Wu, Xinyu Wang and Shanjuan Xie
Minerals 2024, 14(9), 923; https://doi.org/10.3390/min14090923 - 10 Sep 2024
Viewed by 245
Abstract
The issues of the same material with different spectra and the same spectra for different materials pose challenges in hyperspectral rock classification. This paper proposes a multidimensional feature network based on 2-D convolutional neural networks (2-D CNNs) and recurrent neural networks (RNNs) for [...] Read more.
The issues of the same material with different spectra and the same spectra for different materials pose challenges in hyperspectral rock classification. This paper proposes a multidimensional feature network based on 2-D convolutional neural networks (2-D CNNs) and recurrent neural networks (RNNs) for achieving deep combined extraction and fusion of spatial information, such as the rock shape and texture, with spectral information. Experiments are conducted on a hyperspectral rock image dataset obtained by scanning 81 common igneous and metamorphic rock samples using the HySpex hyperspectral sensor imaging system to validate the effectiveness of the proposed network model. The results show that the model achieved an overall classification accuracy of 97.925% and an average classification accuracy of 97.956% on this dataset, surpassing the performances of existing models in the field of rock classification. Full article
Show Figures

Figure 1

Figure 1
<p>Hyspex hyperspectral sensor.</p>
Full article ">Figure 2
<p>(<b>a</b>) The 81 rock samples, (<b>b</b>) spectral curves of the 28 types of rocks, and (<b>c</b>) rock names corresponding to the 28 types of rocks.</p>
Full article ">Figure 2 Cont.
<p>(<b>a</b>) The 81 rock samples, (<b>b</b>) spectral curves of the 28 types of rocks, and (<b>c</b>) rock names corresponding to the 28 types of rocks.</p>
Full article ">Figure 3
<p>(<b>a</b>) Ground truth for the 28 rock classes, and (<b>b</b>) ground truth after morphological processing.</p>
Full article ">Figure 4
<p>Schematic diagram of the traditional 2-D CNN structure.</p>
Full article ">Figure 5
<p>Schematic diagram of the internal structure of the GRU.</p>
Full article ">Figure 6
<p>Comparison of ASD spectral characteristics of different rocks. (<b>a</b>) Comparison of ASD spectral characteristics of Aplite granite and granite gneiss (<b>b</b>) Comparison of ASD spectral characteristics of andesite and diorite.</p>
Full article ">Figure 7
<p>Image data results of potassium feldspar granite samples.</p>
Full article ">Figure 8
<p>Proposed network structure.</p>
Full article ">Figure 9
<p>Classification effects of different algorithms on the rock dataset: (<b>a</b>) Resnet-18, (<b>b</b>) 3-D CNN, (<b>c</b>) Hamida, (<b>d</b>) HybridSN, (<b>e</b>) CAE-SVM, and (<b>f</b>) our model.</p>
Full article ">Figure 10
<p>Comparison of OAs of all methods for different proportions of the training samples.</p>
Full article ">
23 pages, 21056 KiB  
Article
Development and Application of Unmanned Aerial High-Resolution Convex Grating Dispersion Hyperspectral Imager
by Qingsheng Xue, Xinyu Gao, Fengqin Lu, Jun Ma, Junhong Song and Jinfeng Xu
Sensors 2024, 24(17), 5812; https://doi.org/10.3390/s24175812 - 7 Sep 2024
Viewed by 277
Abstract
This study presents the design and development of a high-resolution convex grating dispersion hyperspectral imaging system tailored for unmanned aerial vehicle (UAV) remote sensing applications. The system operates within a spectral range of 400 to 1000 nm, encompassing over 150 channels, and achieves [...] Read more.
This study presents the design and development of a high-resolution convex grating dispersion hyperspectral imaging system tailored for unmanned aerial vehicle (UAV) remote sensing applications. The system operates within a spectral range of 400 to 1000 nm, encompassing over 150 channels, and achieves an average spectral resolution of less than 4 nm. It features a field of view of 30°, a focal length of 20 mm, a compact volume of only 200 mm × 167 mm × 78 mm, and a total weight of less than 1.5 kg. Based on the design specifications, the system was meticulously adjusted, calibrated, and tested. Additionally, custom software for the hyperspectral system was independently developed to facilitate functions such as control parameter adjustments, real-time display, and data preprocessing of the hyperspectral camera. Subsequently, the prototype was integrated onto a drone for remote sensing observations of Spartina alterniflora at Yangkou Beach in Shouguang City, Shandong Province. Various algorithms were employed for data classification and comparison, with support vector machine (SVM) and neural network algorithms demonstrating superior classification accuracy. The experimental results indicate that the UAV-based hyperspectral imaging system exhibits high imaging quality, minimal distortion, excellent resolution, an expansive camera field of view, a broad detection range, high experimental efficiency, and remarkable capabilities for remote sensing detection. Full article
(This article belongs to the Section Remote Sensors)
Show Figures

Figure 1

Figure 1
<p>Optical structure of the front telescope system.</p>
Full article ">Figure 2
<p>Distribution of point columns on the image plane of the telescope system.</p>
Full article ">Figure 3
<p>Optical transfer function curve of the front telescope system.</p>
Full article ">Figure 4
<p>Optical structure of Offner spectral imaging system.</p>
Full article ">Figure 5
<p>MTF curves for different wavelengths of Offner spectral imaging system: (<b>a</b>) 400 nm; (<b>b</b>) 700 nm; (<b>c</b>) 1000 nm.</p>
Full article ">Figure 5 Cont.
<p>MTF curves for different wavelengths of Offner spectral imaging system: (<b>a</b>) 400 nm; (<b>b</b>) 700 nm; (<b>c</b>) 1000 nm.</p>
Full article ">Figure 6
<p>Curve of RMS radius versus wavelength for Offner spectral imaging system.</p>
Full article ">Figure 7
<p>Map of light imprints on the image plane of Offner spectral imaging system.</p>
Full article ">Figure 8
<p>Spectral bending at different wavelengths of Offner spectral imaging system.</p>
Full article ">Figure 9
<p>Spectral band bending for different fields of view of Offner spectral imaging system.</p>
Full article ">Figure 10
<p>System-wide optical structure of the hyperspectral imager.</p>
Full article ">Figure 11
<p>System-wide optical transfer function curve of hyperspectral imager.</p>
Full article ">Figure 12
<p>Overall mechanical structure of the hyperspectral prototype.</p>
Full article ">Figure 13
<p>Results of Gaussian fitting of some characteristic peaks: (<b>a</b>) 696.54 nm wavelength characteristic peak fitting result; (<b>b</b>) 738.40 nm wavelength characteristic peak fitting result; (<b>c</b>) 763.51 nm wavelength characteristic peak fitting result; (<b>d</b>) 772.40 nm wavelength characteristic peak fitting result; (<b>e</b>) 794.82 nm wavelength characteristic peak fitting result; (<b>f</b>) 912.30 nm wavelength characteristic peak fitting result.</p>
Full article ">Figure 14
<p>Hg lamp calibration fitting results.</p>
Full article ">Figure 15
<p>Overall functional block diagram of hyperspectral control software system.</p>
Full article ">Figure 16
<p>Screenshot of Software System Operation Test.</p>
Full article ">Figure 17
<p>Hyperspectral Imager System Outdoor Rotary Scanning Experiment.</p>
Full article ">Figure 18
<p>Monochromatic images at different wavelength bands from the outdoor push-scan of the hyperspectral imaging system: (<b>a</b>) monochromatic image in the 500 nm wavelength band; (<b>b</b>) monochromatic image in the 600 nm wavelength band; (<b>c</b>) monochromatic image in the 700 nm wavelength band; (<b>d</b>) monochromatic image in the 800 nm wavelength band.</p>
Full article ">Figure 18 Cont.
<p>Monochromatic images at different wavelength bands from the outdoor push-scan of the hyperspectral imaging system: (<b>a</b>) monochromatic image in the 500 nm wavelength band; (<b>b</b>) monochromatic image in the 600 nm wavelength band; (<b>c</b>) monochromatic image in the 700 nm wavelength band; (<b>d</b>) monochromatic image in the 800 nm wavelength band.</p>
Full article ">Figure 19
<p>Spectral intensity curves of roofs, walls, and green trees.</p>
Full article ">Figure 20
<p>Components of an unmanned airborne hyperspectral remote sensing system.</p>
Full article ">Figure 21
<p>Photo of the distribution area of <span class="html-italic">Spartina alterniflora</span> at Yangkou Beach.</p>
Full article ">Figure 22
<p>Monochromatic images of <span class="html-italic">Spartina alterniflora</span> captured by unmanned aerial vehicle hyperspectral system in different frequency bands: (<b>a</b>) monochromatic image in the 560 nm wavelength band; (<b>b</b>) monochromatic image in the 600 nm wavelength band; (<b>c</b>) monochromatic image in the 650 nm wavelength band; (<b>d</b>) monochromatic image in the 700 nm wavelength band; (<b>e</b>) monochromatic image in the 750 band; (<b>f</b>) monochromatic image in the 800 nm wavelength band; (<b>g</b>) monochromatic image in the 850 nm wavelength band.</p>
Full article ">Figure 22 Cont.
<p>Monochromatic images of <span class="html-italic">Spartina alterniflora</span> captured by unmanned aerial vehicle hyperspectral system in different frequency bands: (<b>a</b>) monochromatic image in the 560 nm wavelength band; (<b>b</b>) monochromatic image in the 600 nm wavelength band; (<b>c</b>) monochromatic image in the 650 nm wavelength band; (<b>d</b>) monochromatic image in the 700 nm wavelength band; (<b>e</b>) monochromatic image in the 750 band; (<b>f</b>) monochromatic image in the 800 nm wavelength band; (<b>g</b>) monochromatic image in the 850 nm wavelength band.</p>
Full article ">Figure 23
<p>Spectral curves of <span class="html-italic">Spartina alterniflora</span> Loisel, water, and mudflat.</p>
Full article ">Figure 24
<p>Classification results of hyperspectral data of <span class="html-italic">Spartina alterniflora</span> using different classification algorithms; among them, green represents <span class="html-italic">Spartina alterniflora</span>, light blue represents water area, and dark blue represents mudflat. (<b>a</b>) SAM classification results; (<b>b</b>) SID classification results l; (<b>c</b>) SVM classification results; (<b>d</b>) BPNN classification results.</p>
Full article ">
19 pages, 12043 KiB  
Article
Collection of a Hyperspectral Atmospheric Cloud Dataset and Enhancing Pixel Classification through Patch-Origin Embedding
by Hua Yan, Rachel Zheng, Shivaji Mallela, Randy Russell and Olcay Kursun
Remote Sens. 2024, 16(17), 3315; https://doi.org/10.3390/rs16173315 - 6 Sep 2024
Viewed by 350
Abstract
Hyperspectral cameras collect detailed spectral information at each image pixel, contributing to the identification of image features. The rich spectral content of hyperspectral imagery has led to its application in diverse fields of study. This study focused on cloud classification using a dataset [...] Read more.
Hyperspectral cameras collect detailed spectral information at each image pixel, contributing to the identification of image features. The rich spectral content of hyperspectral imagery has led to its application in diverse fields of study. This study focused on cloud classification using a dataset of hyperspectral sky images captured by a Resonon PIKA XC2 camera. The camera records images using 462 spectral bands, ranging from 400 to 1000 nm, with a spectral resolution of 1.9 nm. Our preliminary/unlabeled dataset comprised 33 parent hyperspectral images (HSI), each a substantial unlabeled image measuring 4402-by-1600 pixels. With the meteorological expertise within our team, we manually labeled pixels by extracting 10 to 20 sample patches from each parent image, each patch consisting of a 50-by-50 pixel field. This process yielded a collection of 444 patches, each categorically labeled into one of seven cloud and sky condition categories. To embed the inherent data structure while classifying individual pixels, we introduced an innovative technique to boost classification accuracy by incorporating patch-specific information into each pixel’s feature vector. The posterior probabilities generated by these classifiers, which capture the unique attributes of each patch, were subsequently concatenated with the pixel’s original spectral data to form an augmented feature vector. We then applied a final classifier to map the augmented vectors to the seven cloud/sky categories. The results compared favorably to the baseline model devoid of patch-origin embedding, showing that incorporating the spatial context along with the spectral information inherent in hyperspectral images enhances the classification accuracy in hyperspectral cloud classification. The dataset is available on IEEE DataPort. Full article
(This article belongs to the Special Issue Deep Learning for Remote Sensing and Geodata)
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>Working principle of Resonon Hyperspectral Imager (inspired from [<a href="#B11-remotesensing-16-03315" class="html-bibr">11</a>]).</p>
Full article ">Figure 2
<p>Resonon Pika XC2 camera mounted on a tilt head and attached to a rotational stage that captures sky images covering a 90-degree range in azimuth [<a href="#B12-remotesensing-16-03315" class="html-bibr">12</a>].</p>
Full article ">Figure 3
<p>A sample parent image with some sample patches marked in red squares.</p>
Full article ">Figure 4
<p>Sample patch examples for each cloud/sky category. (<b>a</b>) Dense dark cumuliform clouds (c01). (<b>b</b>) Dense bright cumuliform clouds (c02). (<b>c</b>) Semi-transparent cumuliform clouds (c03). (<b>d</b>) Dense cirroform clouds (c04). (<b>e</b>) Semi-transparent cirroform clouds (c05). (<b>f</b>) Low aerosol clear sky (c06). (<b>g</b>) Moderate/high aerosol clear sky (c07).</p>
Full article ">Figure 5
<p>Spectra of three exemplary pixels obtained from three separate cumuliform cloud patches, each with 462 bands.</p>
Full article ">Figure 6
<p>Normalized version of the cumuliform spectra of the three pixels and the class average of the normalized spectra.</p>
Full article ">Figure 7
<p>Spectra of three exemplary pixels obtained from three separate cirroform cloud patches, each with 462 bands.</p>
Full article ">Figure 8
<p>Normalized versions of the cirroform spectra of the three pixels and the class average of the normalized spectra.</p>
Full article ">Figure 9
<p>Spectra of three exemplary pixels obtained from three separate clear-sky patches, each with 462 bands.</p>
Full article ">Figure 10
<p>Normalized version of the clear-sky spectra of the three pixels and the class average of the normalized spectra.</p>
Full article ">Figure 11
<p>Notations used for the origin of a parent image and the location and size of a patch in the parent image.</p>
Full article ">Figure 12
<p>Parent image file naming convention of the dataset [<a href="#B10-remotesensing-16-03315" class="html-bibr">10</a>].</p>
Full article ">Figure 13
<p>Patch image file naming convention of the dataset [<a href="#B10-remotesensing-16-03315" class="html-bibr">10</a>].</p>
Full article ">Figure 14
<p>CNN architecture used for patch classification using the RGB renders.</p>
Full article ">Figure 15
<p>CNN architecture used for feature extraction. Outputs from the network’s GlobalMaxPooling layer serve as features to downstream classifiers, either LR or RF.</p>
Full article ">Figure 16
<p>Classification results on sample parent images (from [<a href="#B21-remotesensing-16-03315" class="html-bibr">21</a>]), which are large images not included in the training dataset. The results demonstrate the performance of the classification model on new, unseen data.</p>
Full article ">
16 pages, 5675 KiB  
Article
Utilization of Machine Learning and Hyperspectral Imaging Technologies for Classifying Coated Maize Seed Vigor: A Case Study on the Assessment of Seed DNA Repair Capability
by Kris Wonggasem, Papis Wongchaisuwat, Pongsan Chakranon and Damrongvudhi Onwimol
Agronomy 2024, 14(9), 1991; https://doi.org/10.3390/agronomy14091991 - 2 Sep 2024
Viewed by 374
Abstract
The conventional evaluation of maize seed vigor is a time-consuming and labor-intensive process. By contrast, this study introduces an automated, nondestructive framework for classifying maize seed vigor with different seed DNA repair capabilities using hyperspectral images. The selection of coated maize seeds for [...] Read more.
The conventional evaluation of maize seed vigor is a time-consuming and labor-intensive process. By contrast, this study introduces an automated, nondestructive framework for classifying maize seed vigor with different seed DNA repair capabilities using hyperspectral images. The selection of coated maize seeds for our case study also aligned well with practical applications. To ensure the accuracy and reliability of the results, rigorous data preprocessing steps were implemented to extract high-quality information from raw spectral data obtained from the hyperspectral images. In particular, commonly used pretreatment methods were explored. Instead of analyzing all the wavelengths of spectral data, a competitive adaptive reweighted sampling method was used to select more informative wavelengths, optimizing analysis efficiency. Furthermore, this study leveraged machine learning models, enriched through oversampling techniques to address data imbalance at the seed level. The results obtained using a support vector machine with enhanced techniques demonstrated promising results with 100% sensitivity, 96.91% specificity, and a 0.9807 Matthews correlation coefficient (MCC). Thus, this study highlighted the effectiveness of hyperspectral imaging and machine learning in modern seed assessment practices. By introducing a seed vigor classification system that can even accommodate coated seeds, this study offers a potential pathway for empowering seed producers in practical, real-world applications. Full article
Show Figures

Figure 1

Figure 1
<p>Physical map of hyperspectral imaging system. The hyperspectral instrument for image acquisition (<b>A</b>) and maize seed characteristics employed for the acquisition of spectral data (<b>B</b>).</p>
Full article ">Figure 2
<p>ROI process with spectral extraction.</p>
Full article ">Figure 3
<p>Seed germination of 120 coated maize seed samples was obtained from different lots produced in Thailand over 3 years. The germinator was set to 25 °C. Error bars denote confidence intervals (<span class="html-italic">n</span> = 4; <span class="html-italic">p</span> &lt; 0.05); missing error bars indicate ranges smaller than the symbols.</p>
Full article ">Figure 4
<p>Illustrations of the radicle emergence characteristics of seeds with high vigor compared with those with low vigor at each time point.</p>
Full article ">Figure 5
<p>Four-parameter hill function-fitted cumulative radicle emergence curve. Blue and red represent high- and low-vigor seed lots, respectively (<b>A</b>). Classification of seed quality for 120 maize seed lots according to K-means clustering using the following metrics: area under the fitted curve for radicle emergence, uniformity of radicle emergence, mean radicle emergence times, and radicle emergence speed employed as the ground truth labels (<b>B</b>).</p>
Full article ">Figure 6
<p>MCC obtained from ELDA (<b>top</b>) and SVM (<b>bottom</b>) using varied oversampling parameters based on ADASYN (<b>left</b>) and BorderlineSMOTE (<b>right</b>) with original spectral, SNV, and MSC pretreatment analysis.</p>
Full article ">Figure 7
<p>Confusion matrix (<b>A</b>) and ROC curve (<b>B</b>) of the superior model.</p>
Full article ">
22 pages, 9228 KiB  
Article
Cross-Hopping Graph Networks for Hyperspectral–High Spatial Resolution (H2) Image Classification
by Tao Chen, Tingting Wang, Huayue Chen, Bochuan Zheng and Wu Deng
Remote Sens. 2024, 16(17), 3155; https://doi.org/10.3390/rs16173155 - 27 Aug 2024
Viewed by 457
Abstract
As we take stock of the contemporary issue, remote sensing images are gradually advancing towards hyperspectral–high spatial resolution (H2) double-high images. However, high resolution produces serious spatial heterogeneity and spectral variability while improving image resolution, which increases the difficulty of feature [...] Read more.
As we take stock of the contemporary issue, remote sensing images are gradually advancing towards hyperspectral–high spatial resolution (H2) double-high images. However, high resolution produces serious spatial heterogeneity and spectral variability while improving image resolution, which increases the difficulty of feature recognition. So as to make the best of spectral and spatial features under an insufficient number of marking samples, we would like to achieve effective recognition and accurate classification of features in H2 images. In this paper, a cross-hop graph network for H2 image classification(H2-CHGN) is proposed. It is a two-branch network for deep feature extraction geared towards H2 images, consisting of a cross-hop graph attention network (CGAT) and a multiscale convolutional neural network (MCNN): the CGAT branch utilizes the superpixel information of H2 images to filter samples with high spatial relevance and designate them as the samples to be classified, then utilizes the cross-hop graph and attention mechanism to broaden the range of graph convolution to obtain more representative global features. As another branch, the MCNN uses dual convolutional kernels to extract features and fuse them at various scales while attaining pixel-level multi-scale local features by parallel cross connecting. Finally, the dual-channel attention mechanism is utilized for fusion to make image elements more prominent. This experiment on the classical dataset (Pavia University) and double-high (H2) datasets (WHU-Hi-LongKou and WHU-Hi-HongHu) shows that the H2-CHGN can be efficiently and competently used in H2 image classification. In detail, experimental results showcase superior performance, outpacing state-of-the-art methods by 0.75–2.16% in overall accuracy. Full article
(This article belongs to the Special Issue Deep Learning for Spectral-Spatial Hyperspectral Image Classification)
Show Figures

Figure 1

Figure 1
<p>The framework of H<sup>2</sup>-CHGN model for H<sup>2</sup> image classification.</p>
Full article ">Figure 2
<p>Superpixel and pixel feature conversion process.</p>
Full article ">Figure 3
<p>The procedure for k-hop matrices.</p>
Full article ">Figure 4
<p>Cross-hop graph attention module: (<b>a</b>) pyramid structure by cross-connect feature and (<b>b</b>) graph attention mechanism.</p>
Full article ">Figure 5
<p>The structure of ConvBlock.</p>
Full article ">Figure 6
<p>Convolutional block attention network module: (<b>a</b>) channel attention module and (<b>b</b>) spatial attention module.</p>
Full article ">Figure 7
<p>Dual-channel attention fusion module.</p>
Full article ">Figure 8
<p>OAs under a different number of heads and epochs: (<b>a</b>) Pavia University; (<b>b</b>) WHU-Hi-LongKou; and (<b>c</b>) WHU-Hi-HongHu.</p>
Full article ">Figure 9
<p>Mean color visualization of superpixel segmented regions on different datasets: (<b>a</b>–<b>d</b>): Pavia University; (<b>e</b>–<b>h</b>): WHU-Hi-LongKou; and (<b>i</b>–<b>l</b>): WHU-Hi-HongHu.</p>
Full article ">Figure 9 Cont.
<p>Mean color visualization of superpixel segmented regions on different datasets: (<b>a</b>–<b>d</b>): Pavia University; (<b>e</b>–<b>h</b>): WHU-Hi-LongKou; and (<b>i</b>–<b>l</b>): WHU-Hi-HongHu.</p>
Full article ">Figure 10
<p>Effect of different superpixel segmentation scales on classification accuracy.</p>
Full article ">Figure 11
<p>Effect of different cross-hopping methods on classification accuracy.</p>
Full article ">Figure 12
<p>Classification maps for the Pavia University dataset: (<b>a</b>) False-color image; (<b>b</b>) ground truth; (<b>c</b>) SVM (OA = 79.54%); (<b>d</b>) CEGCN (OA = 97.81%); (<b>e</b>) SGML (OA = 94.30%); (<b>f</b>) WFCG (OA = 97.53%); (<b>g</b>) MSSG-UNet (OA = 98.52%); (<b>h</b>) MS-RPNet (OA = 96.96%); (<b>i</b>) AMGCFN (OA = 98.24%); and (<b>j</b>) H<sup>2</sup>-CHGN (OA = 99.24%).</p>
Full article ">Figure 13
<p>Classification maps for the WHU-Hi-LongKou dataset: (<b>a</b>) False-color image; (<b>b</b>) ground truth; (<b>c</b>) SVM (OA = 92.88%); (<b>d</b>) CEGCN (OA = 98.72%); (<b>e</b>) SGML (OA = 96.03%); (<b>f</b>) WFCG (OA = 98.29%); (<b>g</b>) MSSG-UNet (OA = 98.56%); (<b>h</b>) MS-RPNet (OA = 97.17%); (<b>i</b>) AMGCFN (OA = 98.44%); and (<b>j</b>) H<sup>2</sup>-CHGN (OA = 99.19%).</p>
Full article ">Figure 14
<p>Classification maps for the WHU-Hi-HongHu dataset: (<b>a</b>) False-color image; (<b>b</b>) Ground truth; (<b>c</b>) SVM (OA = 66.34%); (<b>d</b>) CEGCN (OA = 94.01%); (<b>e</b>) SGML (OA = 92.51%); (<b>f</b>) WFCG (OA = 93.98%); (<b>g</b>) MSSG-UNet (OA = 93.73%); (<b>h</b>) MS-RPNet (OA = 93.56%); (<b>i</b>) AMGCFN (OA = 94.44%); (<b>j</b>) H2-CHGN (OA = 96.60%).</p>
Full article ">Figure 15
<p>Effect of different numbers of training samples for the methods: (<b>a</b>) Pavia University; (<b>b</b>) WHU-Hi-LongKou; and (<b>c</b>) WHU-Hi-HongHu.</p>
Full article ">Figure 16
<p>t-SNE results of different methods on three datasets: (<b>a</b>–<b>d</b>) Pavia University; (<b>e</b>–<b>h</b>) WHU-Hi-LongKou; and (<b>i</b>–<b>l</b>) WHU-Hi-HongHu.</p>
Full article ">Figure 16 Cont.
<p>t-SNE results of different methods on three datasets: (<b>a</b>–<b>d</b>) Pavia University; (<b>e</b>–<b>h</b>) WHU-Hi-LongKou; and (<b>i</b>–<b>l</b>) WHU-Hi-HongHu.</p>
Full article ">
26 pages, 5724 KiB  
Article
Spatial Feature Enhancement and Attention-Guided Bidirectional Sequential Spectral Feature Extraction for Hyperspectral Image Classification
by Yi Liu, Shanjiao Jiang, Yijin Liu and Caihong Mu
Remote Sens. 2024, 16(17), 3124; https://doi.org/10.3390/rs16173124 - 24 Aug 2024
Viewed by 538
Abstract
Hyperspectral images have the characteristics of high spectral resolution and low spatial resolution, which will make the extracted features insufficient and lack detailed information about ground objects, thus affecting the accuracy of classification. The numerous spectral bands of hyperspectral images contain rich spectral [...] Read more.
Hyperspectral images have the characteristics of high spectral resolution and low spatial resolution, which will make the extracted features insufficient and lack detailed information about ground objects, thus affecting the accuracy of classification. The numerous spectral bands of hyperspectral images contain rich spectral features but also bring issues of noise and redundancy. To improve the spatial resolution and fully extract spatial and spectral features, this article proposes an improved feature enhancement and extraction model (IFEE) using spatial feature enhancement and attention-guided bidirectional sequential spectral feature extraction for hyperspectral image classification. The adaptive guided filtering is introduced to highlight details and edge features in hyperspectral images. Then, an image enhancement module composed of two-dimensional convolutional neural networks is used to improve the resolution of the image after adaptive guidance filtering and provide a high-resolution image with key features emphasized for the subsequent feature extraction module. The proposed spectral attention mechanism helps to extract more representative spectral features, emphasizing useful information while suppressing the interference of noise. Experimental results show that our method outperforms other comparative methods even with very few training samples. Full article
(This article belongs to the Section Remote Sensing Image Processing)
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>The feature extraction principle of 2D convolution.</p>
Full article ">Figure 2
<p>The architecture of the proposed IFEE. In the spatial feature extraction branch, the adaptive guided filtering process supplements the edge information to the whole HSI input. Then, the filtered image gets improved in spatial resolution by the image enhancement module, and the enhanced HSI patches are sent to the MMFE module to obtain multi-scale spatial features. Spectral features are extracted by the attention-guided bidirectional sequential spectral feature extraction module with two forms of inputs: patch input and vector input. Spatial and spectral features will be concatenated and then classified through Softmax.</p>
Full article ">Figure 3
<p>The images before and after adaptive guided filtering processing on SS dataset. (<b>a</b>) Input image; (<b>b</b>) Output image.</p>
Full article ">Figure 4
<p>The image enhancement module with 2DCNNs. During the pre-processing stage, the input image cube P is first reduced by a certain proportion through bicubic interpolation and obtains a smaller image cube B, which is enlarged and restored to image P+ with the same size as the original input P. The three 2D convolution layers with different sizes of convolution kernels after the pre-processing sequentially enhance the spatial information with the reconstruction of image P+; thus, the output H has higher spatial resolution compared to input image P. We take mean square error (MSE) as the loss function for back propagation and parameter update.</p>
Full article ">Figure 5
<p>The specific structure of BiLSTM and LSTM. In BiLSTM, forward and reverse LSTMs perform spectral feature extraction on the input vector in both front and back directions. Output (1) and Output (2) are the output feature vectors of the reverse processing and the forward processing, respectively, which will be concatenated as the final output feature vector Y. LSTM is mainly composed of the forget gate, input gate, and output gate, which are combinations of the sigmoid and tanh functions. Input h (t − 1) means the output of the last cell, and h (t) is the output we need. C (t−1) adds the memory state of the last cell and is updated to C (t) as an output, and r (t) is the input of the current cell. (<b>a</b>) BiLSTM; (<b>b</b>) LSTM.</p>
Full article ">Figure 6
<p>The structure of the attention mechanism. We add the output vectors after the input patch goes through a two-branch pooling process to combine the global and local spectral information. Then, there are two convolution blocks for the attention weight which will be multiplied by the input vector transformed into a sequence. By synthesizing the pooled sequence after addition and the input vector before and after weight processing, we obtain the output vector for the process of BiLSTM.</p>
Full article ">Figure 7
<p>Classification maps generated by different models on the IP dataset. (<b>a</b>) Ground Truth; (<b>b</b>) SVM; (<b>c</b>) 3DCNN; (<b>d</b>) LSTM; (<b>e</b>) MSCNN; (<b>f</b>) HybridSN; (<b>g</b>) SSDF; (<b>h</b>) SSFTT; (<b>i</b>) IFEE.</p>
Full article ">Figure 8
<p>Classification maps generated by different models on the UP dataset. (<b>a</b>) Ground Truth; (<b>b</b>) SVM; (<b>c</b>) 3DCNN; (<b>d</b>) LSTM; (<b>e</b>) MSCNN; (<b>f</b>) HybridSN; (<b>g</b>) SSDF; (<b>h</b>) SSFTT; (<b>i</b>) IFEE.</p>
Full article ">Figure 9
<p>Classification maps generated by different models on the SS dataset. (<b>a</b>) Ground Truth; (<b>b</b>) SVM; (<b>c</b>) 3DCNN; (<b>d</b>) LSTM; (<b>e</b>) MSCNN; (<b>f</b>) HybridSN; (<b>g</b>) SSDF; (<b>h</b>) SSFTT; (<b>i</b>) IFEE.</p>
Full article ">Figure 9 Cont.
<p>Classification maps generated by different models on the SS dataset. (<b>a</b>) Ground Truth; (<b>b</b>) SVM; (<b>c</b>) 3DCNN; (<b>d</b>) LSTM; (<b>e</b>) MSCNN; (<b>f</b>) HybridSN; (<b>g</b>) SSDF; (<b>h</b>) SSFTT; (<b>i</b>) IFEE.</p>
Full article ">Figure 10
<p>The 2D gray images before and after enhancement processing on IP, UP, and SS datasets. (<b>a</b>) IP before; (<b>b</b>) UP before; (<b>c</b>) SS before; (<b>d</b>) IP after; (<b>e</b>) UP after; (<b>f</b>) SS after.</p>
Full article ">Figure 11
<p>The curves of OA values of IFEE and other methods using two different ways of selecting training samples. (<b>a</b>–<b>c</b>) Curves with different proportions of training samples in three datasets. (<b>d</b>–<b>f</b>) Curves with a different number of training samples in each class in three datasets. (<b>a</b>) Different proportions of training samples in IP. (<b>b</b>) Different proportions of training samples in UP. (<b>c</b>) Different proportions of training samples in SS. (<b>d</b>) Different numbers of training samples in each class in IP. (<b>e</b>) A different number of training samples in each class in UP. (<b>f</b>) Different numbers of training samples in each class in SS.</p>
Full article ">
28 pages, 24617 KiB  
Article
Noise-Disruption-Inspired Neural Architecture Search with Spatial–Spectral Attention for Hyperspectral Image Classification
by Aili Wang, Kang Zhang, Haibin Wu, Shiyu Dai, Yuji Iwahori and Xiaoyu Yu
Remote Sens. 2024, 16(17), 3123; https://doi.org/10.3390/rs16173123 - 24 Aug 2024
Viewed by 606
Abstract
In view of the complexity and diversity of hyperspectral images (HSIs), the classification task has been a major challenge in the field of remote sensing image processing. Hyperspectral classification (HSIC) methods based on neural architecture search (NAS) is a current attractive frontier that [...] Read more.
In view of the complexity and diversity of hyperspectral images (HSIs), the classification task has been a major challenge in the field of remote sensing image processing. Hyperspectral classification (HSIC) methods based on neural architecture search (NAS) is a current attractive frontier that not only automatically searches for neural network architectures best suited to the characteristics of HSI data, but also avoids the possible limitations of manual design of neural networks when dealing with new classification tasks. However, the existing NAS-based HSIC methods have the following limitations: (1) the search space lacks efficient convolution operators that can fully extract discriminative spatial–spectral features, and (2) NAS based on traditional differentiable architecture search (DARTS) has performance collapse caused by unfair competition. To overcome these limitations, we proposed a neural architecture search method with receptive field spatial–spectral attention (RFSS-NAS), which is specifically designed to automatically search the optimal architecture for HSIC. Considering the core needs of the model in extracting more discriminative spatial–spectral features, we designed a novel and efficient attention search space. The core component of this innovative space is the receptive field spatial–spectral attention convolution operator, which is capable of precisely focusing on the critical information in the image, thus greatly enhancing the quality of feature extraction. Meanwhile, for the purpose of solving the unfair competition issue in the traditional differentiable architecture search (DARTS) strategy, we skillfully introduce the Noisy-DARTS strategy. The strategy ensures the fairness and efficiency of the search process and effectively avoids the risk of performance crash. In addition, to further improve the robustness of the model and ability to recognize difficult-to-classify samples, we proposed a fusion loss function by combining the advantages of the label smoothing loss and the polynomial expansion perspective loss function, which not only smooths the label distribution and reduces the risk of overfitting, but also effectively handles those difficult-to-classify samples, thus improving the overall classification accuracy. Experiments on three public datasets fully validate the superior performance of RFSS-NAS. Full article
(This article belongs to the Special Issue Recent Advances in the Processing of Hyperspectral Images)
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>The proposed framework of RFSS-NAS for HSIC.</p>
Full article ">Figure 2
<p>Receptive field spatial–spectral attention separable convolution operators (K = 3, 5, 7).</p>
Full article ">Figure 3
<p>The structure of Fused_MBConv.</p>
Full article ">Figure 4
<p>The Noisy-DARTS search process: (<b>a</b>) building candidate operations between nodes; and (<b>b</b>) the final architecture of topological architecture search.</p>
Full article ">Figure 5
<p>The classification results of KSC dataset. (<b>a</b>) Ground-truth map. (<b>b</b>) RBF-SVM. (<b>c</b>) CNN. (<b>d</b>) PyResNet. (<b>e</b>) SSRN. (<b>f</b>) 3-D AT-CNN. (<b>g</b>) HNAS. (<b>h</b>) LMSS-NAS. (<b>i</b>) RFSS-NAS.(The red box represents the selected area enlarged for comparison.)</p>
Full article ">Figure 5 Cont.
<p>The classification results of KSC dataset. (<b>a</b>) Ground-truth map. (<b>b</b>) RBF-SVM. (<b>c</b>) CNN. (<b>d</b>) PyResNet. (<b>e</b>) SSRN. (<b>f</b>) 3-D AT-CNN. (<b>g</b>) HNAS. (<b>h</b>) LMSS-NAS. (<b>i</b>) RFSS-NAS.(The red box represents the selected area enlarged for comparison.)</p>
Full article ">Figure 6
<p>The classification results of PU dataset. (<b>a</b>) Ground-truth map. (<b>b</b>) RBF-SVM. (<b>c</b>) CNN. (<b>d</b>) PyResNet. (<b>e</b>) SSRN. (<b>f</b>) 3-D AT-CNN. (<b>g</b>) HNAS. (<b>h</b>) LMSS-NAS. (<b>i</b>) RFSS-NAS.</p>
Full article ">Figure 7
<p>The classification results of HU dataset. (<b>a</b>) Ground-truth map. (<b>b</b>) RBF-SVM. (<b>c</b>) CNN. (<b>d</b>) PyResNet. (<b>e</b>) SSRN. (<b>f</b>) 3-D AT-CNN. (<b>g</b>) HNAS. (<b>h</b>) LMSS-NAS. (<b>i</b>) RFSS-NAS. (The red and blue boxed represent the selected area enlarged for comparison.)</p>
Full article ">Figure 7 Cont.
<p>The classification results of HU dataset. (<b>a</b>) Ground-truth map. (<b>b</b>) RBF-SVM. (<b>c</b>) CNN. (<b>d</b>) PyResNet. (<b>e</b>) SSRN. (<b>f</b>) 3-D AT-CNN. (<b>g</b>) HNAS. (<b>h</b>) LMSS-NAS. (<b>i</b>) RFSS-NAS. (The red and blue boxed represent the selected area enlarged for comparison.)</p>
Full article ">Figure 8
<p>The searched cell architectures of KSC dataset. (<b>a</b>) Normal cell. (<b>b</b>) Reduce cell.</p>
Full article ">Figure 9
<p>The searched cell architectures of PU dataset. (<b>a</b>) Normal cell. (<b>b</b>) Reduce cell.</p>
Full article ">Figure 10
<p>The searched cell architectures of HU dataset. (<b>a</b>) Normal cell. (<b>b</b>) Reduce cell.</p>
Full article ">Figure 11
<p>The researched normal cell of different architectures: (<b>a</b>) RFSS-NA; (<b>b</b>) RFSS-NAS-Rop; (<b>c</b>) RFSS-NAS-Rtopo; and (<b>d</b>) RFSS-NAS-Rbop.</p>
Full article ">Figure 12
<p>The researched reduction cell of different architectures: (<b>a</b>) RFSS-NA; (<b>b</b>) RFSS-NAS-Rop; (<b>c</b>) RFSS-NAS-Rtopo; and (<b>d</b>) RFSS-NAS-Rbop.</p>
Full article ">Figure 13
<p>Visualization of the 2D spectral–spatial features in KSC via t-SNE. (<b>a</b>) Without RFSSA. (<b>b</b>) With RFFSA.</p>
Full article ">Figure 14
<p>The confusion matrices of KSC, PU, and HU datasets: (<b>a</b>) KSC; (<b>b</b>) PU; and (<b>c</b>) HU.</p>
Full article ">Figure 15
<p>Learning curves on the KSC dataset. (<b>a</b>) Valid loss vs train loss each epoch. (<b>b</b>) Valid accuracy vs train accuracy each epoch.</p>
Full article ">Figure 16
<p>Learning curves on the PU dataset. (<b>a</b>) Valid loss vs train loss each epoch. (<b>b</b>) Valid accuracy vs train accuracy each epoch.</p>
Full article ">Figure 17
<p>Learning curves on the HU dataset. (<b>a</b>) Valid loss vs train loss each epoch. (<b>b</b>) Valid accuracy vs train accuracy each epoch.</p>
Full article ">
23 pages, 39394 KiB  
Article
Fine-Scale Mangrove Species Classification Based on UAV Multispectral and Hyperspectral Remote Sensing Using Machine Learning
by Yuanzheng Yang, Zhouju Meng, Jiaxing Zu, Wenhua Cai, Jiali Wang, Hongxin Su and Jian Yang
Remote Sens. 2024, 16(16), 3093; https://doi.org/10.3390/rs16163093 - 22 Aug 2024
Viewed by 809
Abstract
Mangrove ecosystems play an irreplaceable role in coastal environments by providing essential ecosystem services. Diverse mangrove species have different functions due to their morphological and physiological characteristics. A precise spatial distribution map of mangrove species is therefore crucial for biodiversity maintenance and environmental [...] Read more.
Mangrove ecosystems play an irreplaceable role in coastal environments by providing essential ecosystem services. Diverse mangrove species have different functions due to their morphological and physiological characteristics. A precise spatial distribution map of mangrove species is therefore crucial for biodiversity maintenance and environmental conservation of coastal ecosystems. Traditional satellite data are limited in fine-scale mangrove species classification due to low spatial resolution and less spectral information. This study employed unmanned aerial vehicle (UAV) technology to acquire high-resolution multispectral and hyperspectral mangrove forest imagery in Guangxi, China. We leveraged advanced algorithms, including RFE-RF for feature selection and machine learning models (Adaptive Boosting (AdaBoost), eXtreme Gradient Boosting (XGBoost), Random Forest (RF), and Light Gradient Boosting Machine (LightGBM)), to achieve mangrove species mapping with high classification accuracy. The study assessed the classification performance of these four machine learning models for two types of image data (UAV multispectral and hyperspectral imagery), respectively. The results demonstrated that hyperspectral imagery had superiority over multispectral data by offering enhanced noise reduction and classification performance. Hyperspectral imagery produced mangrove species classification with overall accuracy (OA) higher than 91% across the four machine learning models. LightGBM achieved the highest OA of 97.15% and kappa coefficient (Kappa) of 0.97 based on hyperspectral imagery. Dimensionality reduction and feature extraction techniques were effectively applied to the UAV data, with vegetation indices proving to be particularly valuable for species classification. The present research underscored the effectiveness of UAV hyperspectral images using machine learning models for fine-scale mangrove species classification. This approach has the potential to significantly improve ecological management and conservation strategies, providing a robust framework for monitoring and safeguarding these essential coastal habitats. Full article
Show Figures

Figure 1

Figure 1
<p>Study area and UAV-based visible image ((<b>A</b>): Yingluo Bay, (<b>B</b>): Pearl Bay).</p>
Full article ">Figure 2
<p>Workflow diagram illustrating the methodology of this study.</p>
Full article ">Figure 3
<p>Mangrove species classification comparison of user’s and producer’s accuracies obtained by four learning models based on multi- and hyper-spectral images in Yingluo Bay.</p>
Full article ">Figure 4
<p>Mangrove species classification comparison of user’s and producer’s accuracies obtained by LightGBM learning model based on the multi- and hyper-spectral image in Pearl Bay.</p>
Full article ">Figure 5
<p>The mangrove species classification maps using four learning models (LightGBM, RF, XGBoost, and AdaBoost) based on UAV multispectral image (<b>a</b>–<b>d</b>) and hyperspectral image (<b>e</b>–<b>h</b>), respectively, in Yingluo Bay.</p>
Full article ">Figure 6
<p>The UAV visual image covering Yingluo Bay and three subsets (<b>A</b>–<b>C</b>) of the UAV multispectral and hyperspectral image classification results based on the LightGBM learning model.</p>
Full article ">Figure 7
<p>The mangrove species classification maps using the LightGBM learning model based on UAV multispectral image (<b>a</b>) and hyperspectral image (<b>b</b>) in Pearl Bay.</p>
Full article ">Figure 8
<p>The UAV visual image covering Pearl Bay and three subsets (<b>A</b>–<b>C</b>) of the UAV multispectral and hyperspectral image classification results using LightGBM learning model.</p>
Full article ">Figure A1
<p>Normalized confusion matrices of mangrove species classification using four learning models (AdaBoost, XGboost, RF, and LightGBM) based on UAV multi- and hyper-spectral images in Yingluo Bay.</p>
Full article ">
24 pages, 4633 KiB  
Article
Multimodal Semantic Collaborative Classification for Hyperspectral Images and LiDAR Data
by Aili Wang, Shiyu Dai, Haibin Wu and Yuji Iwahori
Remote Sens. 2024, 16(16), 3082; https://doi.org/10.3390/rs16163082 - 21 Aug 2024
Viewed by 714
Abstract
Although the collaborative use of hyperspectral images (HSIs) and LiDAR data in land cover classification tasks has demonstrated significant importance and potential, several challenges remain. Notably, the heterogeneity in cross-modal information integration presents a major obstacle. Furthermore, most existing research relies heavily on [...] Read more.
Although the collaborative use of hyperspectral images (HSIs) and LiDAR data in land cover classification tasks has demonstrated significant importance and potential, several challenges remain. Notably, the heterogeneity in cross-modal information integration presents a major obstacle. Furthermore, most existing research relies heavily on category names, neglecting the rich contextual information from language descriptions. Visual-language pretraining (VLP) has achieved notable success in image recognition within natural domains by using multimodal information to enhance training efficiency and effectiveness. VLP has also shown great potential for land cover classification in remote sensing. This paper introduces a dual-sensor multimodal semantic collaborative classification network (DSMSC2N). It uses large language models (LLMs) in an instruction-driven manner to generate land cover category descriptions enriched with domain-specific knowledge in remote sensing. This approach aims to guide the model to accurately focus on and extract key features. Simultaneously, we integrate and optimize the complementary relationship between HSI and LiDAR data, enhancing the separability of land cover categories and improving classification accuracy. We conduct comprehensive experiments on benchmark datasets like Houston 2013, Trento, and MUUFL Gulfport, validating DSMSC2N’s effectiveness compared to various baseline methods. Full article
(This article belongs to the Special Issue Recent Advances in the Processing of Hyperspectral Images)
Show Figures

Figure 1

Figure 1
<p>An overview of the proposed DSMSC<sup>2</sup>N.</p>
Full article ">Figure 2
<p>Workflow for automated construction of a high-dimensional spectral class descriptor collection.</p>
Full article ">Figure 3
<p>Graphical representation of ModaUnion encoder.</p>
Full article ">Figure 4
<p>The mechanism of clustering.</p>
Full article ">Figure 5
<p>The visualization of the Houston 2013 dataset. (<b>a</b>) Pseudo color map of an HSI. (<b>b</b>) DSM of LiDAR. (<b>c</b>) Training sample map. (<b>d</b>) Testing sample map.</p>
Full article ">Figure 6
<p>The visualization of the Trento dataset. (<b>a</b>) Pseudo color map of an HSI. (<b>b</b>) DSM of LiDAR. (<b>c</b>) Training sample map. (<b>d</b>) Testing sample map.</p>
Full article ">Figure 7
<p>The visualization of the MUUFL Gulfport dataset. (<b>a</b>) Pseudo color map of an HSI. (<b>b</b>) DSM of LiDAR. (<b>c</b>) Training sample map. (<b>d</b>) Testing sample map.</p>
Full article ">Figure 8
<p>T-SNE visualization of loss functions on Trento. (<b>a</b>) CE; (<b>b</b>) without HTBCL; (<b>c</b>) all.</p>
Full article ">Figure 9
<p>Classification maps of Houston 2013. (<b>a</b>) Ground-truth map; (<b>b</b>) two-branch. (<b>c</b>) EndNet; (<b>d</b>) MDL-Middle; (<b>e</b>) MAHiDFNet; (<b>f</b>) FusAtNet; (<b>g</b>) CALC; (<b>h</b>) SepG-ResNet50; (<b>i</b>) DSMSC<sup>2</sup>N.</p>
Full article ">Figure 10
<p>Classification maps of Trento. (<b>a</b>) Ground-truth map; (<b>b</b>) two-branch; (<b>c</b>) EndNet; (<b>d</b>) MDL-Middle; (<b>e</b>) MAHiDFNet; (<b>f</b>) FusAtNet; (<b>g</b>) CALC; (<b>h</b>) SepG-ResNet50; (<b>i</b>) DSMSC<sup>2</sup>N.</p>
Full article ">Figure 11
<p>Classification maps of MUUFL Gulfport. (<b>a</b>) Ground-truth map; (<b>b</b>) two-branch; (<b>c</b>) EndNet; (<b>d</b>) MDL-Middle; (<b>e</b>) MAHiDFNet; (<b>f</b>) FusAtNet; (<b>g</b>) CALC; (<b>h</b>) SepG-ResNet50; (<b>i</b>) DSMSC<sup>2</sup>N.</p>
Full article ">
15 pages, 2400 KiB  
Article
SF-ICNN: Spectral–Fractal Iterative Convolutional Neural Network for Classification of Hyperspectral Images
by Behnam Asghari Beirami, Mehran Alizadeh Pirbasti and Vahid Akbari
Appl. Sci. 2024, 14(16), 7361; https://doi.org/10.3390/app14167361 - 21 Aug 2024
Viewed by 454
Abstract
One primary concern in the field of remote-sensing image processing is the precise classification of hyperspectral images (HSIs). Lately, deep-learning models have demonstrated cutting-edge results in HSI classification. Despite this, researchers continue to study and propose simpler, more robust models. This study presents [...] Read more.
One primary concern in the field of remote-sensing image processing is the precise classification of hyperspectral images (HSIs). Lately, deep-learning models have demonstrated cutting-edge results in HSI classification. Despite this, researchers continue to study and propose simpler, more robust models. This study presents a novel deep-learning approach, the iterative convolutional neural network (ICNN), which combines spectral–fractal features and classifier probability maps iteratively, aiming to enhance the HSI classification accuracy. Experiments are conducted to prove the accuracy enhancement of the proposed method using HSI benchmark datasets of Indian pine (IP) and the University of Pavia (PU) to evaluate the performance of the proposed technique. The final results show that the proposed approach reaches overall accuracies of 99.16% and 95.5% on the IP and PU datasets, respectively, which are better than some basic methods. Additionally, the end findings demonstrate that greater accuracy levels might be achieved using a primary CNN network that employs the iteration loop than with certain current state-of-the-art spatial–spectral HSI classification techniques. Full article
(This article belongs to the Special Issue Hyperspectral Image: Research and Applications)
Show Figures

Figure 1

Figure 1
<p>The suggested SF-ICNN classification method’s flowchart. The proposed method consists of several key steps: spectral–fractal feature generation, CNN classification, and iterative integration of PMs into the feature data cube to improve the classification accuracy.</p>
Full article ">Figure 2
<p>The CNN’s main layers.</p>
Full article ">Figure 3
<p>The IP dataset.</p>
Full article ">Figure 4
<p>The PU dataset.</p>
Full article ">Figure 5
<p>GT and other techniques’ classification results for the IP HSI dataset. (<b>a</b>) GT. (<b>b</b>) S-SVM. (<b>c</b>) SF-SVM. (<b>d</b>) S-CNN. (<b>e</b>) SF-CNN. (<b>f</b>) Suggested SF-ICNN.</p>
Full article ">Figure 6
<p>GT and other techniques’ classification results for the PU HSI dataset. (<b>a</b>) GT. (<b>b</b>) S-SVM. (<b>c</b>) SF-SVM. (<b>d</b>) S-CNN. (<b>e</b>) SF-CNN. (<b>f</b>) Suggested SF-ICNN.</p>
Full article ">Figure 7
<p>Variation of the suggested method’s OAs according to the network’s iteration number on validation samples.</p>
Full article ">Figure 8
<p>Variation of the OAs of the proposed SF-ICNN method in relation to the #training: (<b>a</b>) IP, and (<b>b</b>) PU.</p>
Full article ">
22 pages, 16731 KiB  
Article
Advanced Global Prototypical Segmentation Framework for Few-Shot Hyperspectral Image Classification
by Kunming Xia, Guowu Yuan, Mengen Xia, Xiaosen Li, Jinkang Gui and Hao Zhou
Sensors 2024, 24(16), 5386; https://doi.org/10.3390/s24165386 - 21 Aug 2024
Viewed by 567
Abstract
With the advancement of deep learning, related networks have shown strong performance for Hyperspectral Image (HSI) classification. However, these methods face two main challenges in HSI classification: (1) the inability to capture global information of HSI due to the restriction of patch input [...] Read more.
With the advancement of deep learning, related networks have shown strong performance for Hyperspectral Image (HSI) classification. However, these methods face two main challenges in HSI classification: (1) the inability to capture global information of HSI due to the restriction of patch input and (2) insufficient utilization of information from limited labeled samples. To overcome these challenges, we propose an Advanced Global Prototypical Segmentation (AGPS) framework. Within the AGPS framework, we design a patch-free feature extractor segmentation network (SegNet) based on a fully convolutional network (FCN), which processes the entire HSI to capture global information. To enrich the global information extracted by SegNet, we propose a Fusion of Lateral Connection (FLC) structure that fuses the low-level detailed features of the encoder output with the high-level features of the decoder output. Additionally, we propose an Atrous Spatial Pyramid Pooling-Position Attention (ASPP-PA) module to capture multi-scale spatial positional information. Finally, to explore more valuable information from limited labeled samples, we propose an advanced global prototypical representation learning strategy. Building upon the dual constraints of the global prototypical representation learning strategy, we introduce supervised contrastive learning (CL), which optimizes our network with three different constraints. The experimental results of three public datasets demonstrate that our method outperforms the existing state-of-the-art methods. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

Figure 1
<p>The architecture of the AGPS framework.</p>
Full article ">Figure 2
<p>Illustrations of SegNet.</p>
Full article ">Figure 3
<p>ASPP-PA Module.</p>
Full article ">Figure 4
<p>Chikusei data set.</p>
Full article ">Figure 5
<p>IP data set. The black font indicates the category name, and the red font indicates the category serial number.</p>
Full article ">Figure 6
<p>SA data set. The black font indicates the category name, and the red font indicates the category serial number.</p>
Full article ">Figure 7
<p>UP data set. The black font indicates the category name, and the red font indicates the category serial number.</p>
Full article ">Figure 8
<p>Evolution of OA as a function of (<b>a</b>) PCs, (<b>b</b>) <math display="inline"><semantics> <mrow> <mi>α</mi> </mrow> </semantics></math>, where the blue curve represents the SA dataset, the red curve represents the IP dataset, and the yellow curve represents the UP dataset.</p>
Full article ">Figure 9
<p>IP (<b>a</b>) Ground-truth. (<b>b</b>–<b>l</b>) Classification maps for different classifiers. (<b>b</b>) SVM. (<b>c</b>) 3D-CNN. (<b>d</b>) FPGA. (<b>e</b>) DFSL. (<b>f</b>) DFSL + SVM. (<b>g</b>) DFSL + NN. (<b>h</b>) DCFSL. (<b>i</b>) CMTL. (<b>j</b>) S3Net. (<b>k</b>) CRSSNet. (<b>l</b>) AGPS.</p>
Full article ">Figure 10
<p>SA (<b>a</b>) Ground-truth. (<b>b</b>–<b>l</b>) Classification maps for different classifiers. (<b>b</b>) SVM. (<b>c</b>) 3D-CNN. (<b>d</b>) FPGA. (<b>e</b>) DFSL. (<b>f</b>) DFSL + SVM. (<b>g</b>) DFSL + NN. (<b>h</b>) DCFSL. (<b>i</b>) CMTL. (<b>j</b>) S3Net. (<b>k</b>) CRSSNet. (<b>l</b>) AGPS.</p>
Full article ">Figure 11
<p>UP (<b>a</b>) Ground-truth. (<b>b</b>–<b>l</b>) Classification maps for different classifiers. (<b>b</b>) SVM. (<b>c</b>) 3D-CNN. (<b>d</b>) FPGA. (<b>e</b>) DFSL. (<b>f</b>) DFSL + SVM. (<b>g</b>) DFSL + NN. (<b>h</b>) DCFSL. (<b>i</b>) CMTL. (<b>j</b>) S3Net. (<b>k</b>) CRSSNet. (<b>l</b>) AGPS.</p>
Full article ">Figure 12
<p>Feature separability for different methods in the three datasets. (<b>a</b>) S3Net, (<b>b</b>) CRSSNet, and (<b>c</b>) AGPS for IP. (<b>d</b>) S3Net, (<b>e</b>) CRSSNet, and (<b>f</b>) AGPS for SA. (<b>g</b>) S3Net, (<b>h</b>) CRSSNet, and (<b>i</b>) AGPS for UP.</p>
Full article ">Figure 13
<p>Evolution of OA as a function of number of training samples per class. (<b>a</b>) IP (<b>b</b>) SA (<b>c</b>) UP.</p>
Full article ">
21 pages, 1668 KiB  
Article
DCG-Net: Enhanced Hyperspectral Image Classification with Dual-Branch Convolutional Neural Network and Graph Convolutional Neural Network Integration
by Wenkai Zhu, Xueying Sun and Qiang Zhang
Electronics 2024, 13(16), 3271; https://doi.org/10.3390/electronics13163271 - 18 Aug 2024
Viewed by 442
Abstract
In recent years, graph convolutional neural networks (GCNs) and convolutional neural networks (CNNs) have made significant strides in hyperspectral image (HSI) classification. However, existing models often encounter information redundancy and feature mismatch during feature fusion, and they struggle with small-scale refined features. To [...] Read more.
In recent years, graph convolutional neural networks (GCNs) and convolutional neural networks (CNNs) have made significant strides in hyperspectral image (HSI) classification. However, existing models often encounter information redundancy and feature mismatch during feature fusion, and they struggle with small-scale refined features. To address these issues, we propose DCG-Net, an innovative classification network integrating CNN and GCN architectures. Our approach includes the development of a double-branch expanding network (E-Net) to enhance spectral features and efficiently extract high-level features. Additionally, we incorporate a GCN with an attention mechanism to facilitate the integration of multi-space scale superpixel-level and pixel-level features. To further improve feature fusion, we introduce a feature aggregation module (FAM) that adaptively learns channel features, enhancing classification robustness and accuracy. Comprehensive experiments on three widely used datasets show that DCG-Net achieves superior classification results compared to other state-of-the-art methods. Full article
(This article belongs to the Topic Hyperspectral Imaging and Signal Processing)
Show Figures

Figure 1

Figure 1
<p>DCG-Net architecture diagram. Our model is structured into four primary stages: superpixel segmentation, feature extraction, feature aggregation, and feature classification. We employ a dual-branch architecture, where each branch processes both pixel-level and superpixel-level features. The 1st branch retains the original image resolution for feature processing, whereas the 2nd branch processes upscaled feature images to facilitate multi-scale feature analysis.</p>
Full article ">Figure 2
<p>E-Net structure diagram. Both the encoder and decoder use a customized convolution module. E-Net can effectively combine GCN in the dual-branch encoding and decoding processes to achieve the effective combination of different spatial features.</p>
Full article ">Figure 3
<p>Flowchart of superpixel segmentation and graph structure construction. (<b>a</b>) Original hyperspectral image. (<b>b</b>) Hyperspectral image after PCA downscaling, with the 4-connected graph as a feature graph constructed using the pixels of (<b>b</b>) as nodes. (<b>c</b>) Hyperspectral image after superpixel segmentation, with the superpixel graph structure as a feature graph constructed using the superpixels of (<b>c</b>) as nodes. The orange dots represent the graph’s nodes, and the orange dotted lines represent the graph’s edges.</p>
Full article ">Figure 4
<p>Feature aggregation module structure diagram. The module begins with channel feature learning through the channel attention module, then it continues with feature extraction using the customized convolution module.</p>
Full article ">Figure 5
<p>Description of Indian Pines dataset.</p>
Full article ">Figure 6
<p>Description of Kennedy Space Center dataset.</p>
Full article ">Figure 7
<p>Description of the Salinas dataset.</p>
Full article ">Figure 8
<p>Classification maps of different methods for the Indian Pines dataset.</p>
Full article ">Figure 9
<p>Classification maps of different methods for the Kennedy Space Center dataset.</p>
Full article ">Figure 10
<p>Classification maps of different methods for the Salinas dataset.</p>
Full article ">Figure 11
<p>Comparison of classification performance of seven methods with different training set ratios for three datasets.</p>
Full article ">Figure 12
<p>Comparison results of different superpixel numbers on three datasets.</p>
Full article ">Figure 13
<p>Comparison results of different K values on three datasets.</p>
Full article ">
23 pages, 9214 KiB  
Article
DCFF-Net: Deep Context Feature Fusion Network for High-Precision Classification of Hyperspectral Image
by Zhijie Chen, Yu Chen, Yuan Wang, Xiaoyan Wang, Xinsheng Wang and Zhouru Xiang
Remote Sens. 2024, 16(16), 3002; https://doi.org/10.3390/rs16163002 - 15 Aug 2024
Viewed by 541
Abstract
Hyperspectral images (HSI) contain abundant spectral information. Efficient extraction and utilization of this information for image classification remain prominent research topics. Previously, hyperspectral classification techniques primarily relied on statistical attributes and mathematical models of spectral data. Deep learning classification techniques have recently been [...] Read more.
Hyperspectral images (HSI) contain abundant spectral information. Efficient extraction and utilization of this information for image classification remain prominent research topics. Previously, hyperspectral classification techniques primarily relied on statistical attributes and mathematical models of spectral data. Deep learning classification techniques have recently been extensively utilized for hyperspectral data classification, yielding promising outcomes. This study proposes a deep learning approach that uses polarization feature maps for classification. Initially, the polar co-ordinate transformation method was employed to convert the spectral information of all pixels in the image into spectral feature maps. Subsequently, the proposed Deep Context Feature Fusion Network (DCFF-NET) was utilized to classify these feature maps. The model was validated using three open-source hyperspectral datasets: Indian Pines, Pavia University, and Salinas. The experimental results indicated that DCFF-NET achieved excellent classification performance. Experimental results on three public HSI datasets demonstrated that the proposed method accurately recognized different objects with an overall accuracy (OA) of 86.68%, 94.73%, and 95.14% based on the pixel method, and 98.15%, 99.86%, and 99.98% based on the pixel-patch method. Full article
(This article belongs to the Special Issue GeoAI and EO Big Data Driven Advances in Earth Environmental Science)
Show Figures

Figure 1

Figure 1
<p>Feature maps of three datasets. (A representative feature map for each category was selected and displayed from three different datasets.)</p>
Full article ">Figure 2
<p>DCFF-NET network architectures. (Identical processing operations were represented using the same colour background, e.g. orange for convolution operations, green for batch normalisation, blue for activation, light green for pooling, etc.)</p>
Full article ">Figure 3
<p>Hyperspectral dataset: Indian Pines (<b>a</b>,<b>b</b>), Pavia University (<b>c</b>,<b>d</b>), Salinas (<b>e</b>,<b>f</b>).</p>
Full article ">Figure 4
<p>Predicted classification map of 30% samples for training. (<b>a</b>) Three bands false color composite. (<b>b</b>) Ground truth data. (<b>c</b>) NB. (<b>d</b>) KNN. (<b>e</b>) RF. (<b>f</b>) MLP. (<b>g</b>) 1DCNN. (<b>h</b>) SF-Pixel. (<b>i</b>) VGG16. (<b>j</b>) Resnet50. (<b>k</b>) DCFF-NET.</p>
Full article ">Figure 5
<p>Predicted classification map of 30% samples for training. (<b>a</b>) Three bands false color composite. (<b>b</b>) Ground truth data. (<b>c</b>) NB. (<b>d</b>) KNN. (<b>e</b>) RF. (<b>f</b>) MLP. (<b>g</b>) 1DCNN. (<b>h</b>) SF-Pixel. (<b>i</b>) VGG16. (<b>j</b>) Resnet50. (<b>k</b>) DCFF-NET.</p>
Full article ">Figure 6
<p>Predicted classification map of 30% samples for training. (<b>a</b>) Three bands false color composite. (<b>b</b>) Ground truth data. (<b>c</b>) NB. (<b>d</b>) KNN. (<b>e</b>) RF. (<b>f</b>) MLP. (<b>g</b>) 1DCNN. (<b>h</b>) SF-Pixel. (<b>i</b>) VGG16. (<b>j</b>) Resnet50. (<b>k</b>) DCFF-NET.</p>
Full article ">Figure 7
<p>Classification map of IP (<b>A</b>) PU (<b>B</b>) and SA (<b>C</b>) based on patched-based input. (<b>a</b>) False color composite. (<b>b</b>) Ground truth. (<b>c</b>) VGG16. (<b>d</b>) Resnet50. (<b>e</b>) 3-DCNN. (<b>f</b>) HybridSN. (<b>g</b>) A2S2K. (<b>h</b>) SF-Patch. (<b>i</b>) DCFF-NET.</p>
Full article ">Figure 8
<p>Different filling methods. ((<b>a</b>). Feature maps were neither filled inside nor outside; (<b>b</b>). Feature maps were filled internally rather than externally; (<b>c</b>). Feature maps were filled both internally and externally).</p>
Full article ">Figure 9
<p>Three different filling methods training accuracies curve diagrams.</p>
Full article ">Figure 10
<p>Effect of the different numbers of training samples for different methods.</p>
Full article ">
25 pages, 13967 KiB  
Article
Few-Shot Hyperspectral Remote Sensing Image Classification via an Ensemble of Meta-Optimizers with Update Integration
by Tao Hao, Zhihua Zhang and M. James C. Crabbe
Remote Sens. 2024, 16(16), 2988; https://doi.org/10.3390/rs16162988 - 14 Aug 2024
Viewed by 576
Abstract
Hyperspectral images (HSIs) with abundant spectra and high spatial resolution can satisfy the demand for the classification of adjacent homogeneous regions and accurately determine their specific land-cover classes. Due to the potentially large variance within the same class in hyperspectral images, classifying HSIs [...] Read more.
Hyperspectral images (HSIs) with abundant spectra and high spatial resolution can satisfy the demand for the classification of adjacent homogeneous regions and accurately determine their specific land-cover classes. Due to the potentially large variance within the same class in hyperspectral images, classifying HSIs with limited training samples (i.e., few-shot HSI classification) has become especially difficult. To solve this issue without adding training costs, we propose an ensemble of meta-optimizers that were generated one by one through utilizing periodic annealing on the learning rate during the meta-training process. Such a combination of meta-learning and ensemble learning demonstrates a powerful ability to optimize the deep network on few-shot HSI training. In order to further improve the classification performance, we introduced a novel update integration process to determine the most appropriate update for network parameters during the model training process. Compared with popular human-designed optimizers (Adam, AdaGrad, RMSprop, SGD, etc.), our proposed model performed better in convergence speed, final loss value, overall accuracy, average accuracy, and Kappa coefficient on five HSI benchmarks in a few-shot learning setting. Full article
Show Figures

Figure 1

Figure 1
<p>The framework of our proposed method.</p>
Full article ">Figure 2
<p>Loss changes on the PaviaC dataset.</p>
Full article ">Figure 3
<p>Visual classification results on the PaviaC dataset: (<b>a</b>) False-color image. (<b>b</b>) Ground-truth map. Classification maps obtained by (<b>c</b>) Adam (0.9216), (<b>d</b>) AdaGrad (0.9069), (<b>e</b>) RMSprop (0.8965), (<b>f</b>) SGD (0.8253), (<b>g</b>) SGD with momentum (0.9291), (<b>h</b>) LSTM optimizer (0.9342), (<b>i</b>) MOE-A (0.9222), and (<b>j</b>) MOE-U (0.9408).</p>
Full article ">Figure 3 Cont.
<p>Visual classification results on the PaviaC dataset: (<b>a</b>) False-color image. (<b>b</b>) Ground-truth map. Classification maps obtained by (<b>c</b>) Adam (0.9216), (<b>d</b>) AdaGrad (0.9069), (<b>e</b>) RMSprop (0.8965), (<b>f</b>) SGD (0.8253), (<b>g</b>) SGD with momentum (0.9291), (<b>h</b>) LSTM optimizer (0.9342), (<b>i</b>) MOE-A (0.9222), and (<b>j</b>) MOE-U (0.9408).</p>
Full article ">Figure 4
<p>Loss changes on the PaviaU dataset.</p>
Full article ">Figure 5
<p>Visual classification results on the PaviaU dataset: (<b>a</b>) False-color image. (<b>b</b>) Ground-truth map. Classification maps obtained by (<b>c</b>) Adam (0.6918), (<b>d</b>) AdaGrad (0.7050), (<b>e</b>) RMSprop (0.6417), (<b>f</b>) SGD (0.5290), (<b>g</b>) SGD with momentum (0.6475), (<b>h</b>) LSTM optimizer (0.6444), (<b>i</b>) MOE-A (0.6515), and (<b>j</b>) MOE-U (0.6583).</p>
Full article ">Figure 6
<p>Loss changes on the Salinas dataset.</p>
Full article ">Figure 7
<p>Visual classification results on the Salinas dataset: (<b>a</b>) False-color image. (<b>b</b>) Ground-truth map. Classification maps obtained by (<b>c</b>) Adam (0.3873), (<b>d</b>) AdaGrad (0.4571), (<b>e</b>) RMSprop (0.3873), (<b>f</b>) SGD (0.2727), (<b>g</b>) SGD with momentum (0.3885), (<b>h</b>) LSTM optimizer (0.5837), (<b>i</b>) MOE-A (0.6765), and (<b>j</b>) MOE-U (0.6236).</p>
Full article ">Figure 8
<p>Loss changes on the SalinasA dataset.</p>
Full article ">Figure 9
<p>Visual classification results on the SalinasA dataset: (<b>a</b>) False-color image. (<b>b</b>) Ground-truth map. Classification maps obtained by (<b>c</b>) Adam (0.7556), (<b>d</b>) AdaGrad (0.7467), (<b>e</b>) RMSprop (0.7609), (<b>f</b>) SGD (0.6605), (<b>g</b>) SGD with momentum (0.8708), (<b>h</b>) LSTM optimizer (0.9476), (<b>i</b>) MOE-A (0.9139), and (<b>j</b>) MOE-U (0.9277).</p>
Full article ">Figure 10
<p>Loss changes on the PaviaC dataset.</p>
Full article ">Figure 11
<p>Visual classification results on the PaviaC dataset: (<b>a</b>) LSTM optimizer (0.9407), (<b>b</b>) MOE-A (0.9060), and (<b>c</b>) MOE-U (0.9194).</p>
Full article ">Figure 12
<p>Loss changes on the KSC dataset.</p>
Full article ">Figure 13
<p>Visual classification results on the KSC dataset: (<b>a</b>) False-color image. (<b>b</b>) Ground-truth map. Classification maps obtained by (<b>c</b>) Adam (0.4481), (<b>d</b>) AdaGrad (0.3373), (<b>e</b>) RMSprop (0.5296), (<b>f</b>) SGD (0.2810), (<b>g</b>) SGD with momentum (0.4123), (<b>h</b>) LSTM optimizer (0.6173), (<b>i</b>) MOE-A (0.5878), and (<b>j</b>) MOE-U (0.6319).</p>
Full article ">Figure 14
<p>Loss changes on the PaviaC dataset.</p>
Full article ">Figure 15
<p>Visual classification results on the PaviaC dataset: (<b>a</b>) LSTM optimizer (0.9351), (<b>b</b>) MOE-A (0.9401), and (<b>c</b>) MOE-U (0.9423).</p>
Full article ">
Back to TopTop