[go: up one dir, main page]

 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (238)

Search Parameters:
Keywords = pan-sharpening

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
19 pages, 44218 KiB  
Article
Testing the Impact of Pansharpening Using PRISMA Hyperspectral Data: A Case Study Classifying Urban Trees in Naples, Italy
by Miriam Perretta, Gabriele Delogu, Cassandra Funsten, Alessio Patriarca, Eros Caputi and Lorenzo Boccia
Remote Sens. 2024, 16(19), 3730; https://doi.org/10.3390/rs16193730 - 8 Oct 2024
Viewed by 482
Abstract
Urban trees support vital ecological functions and help with the mitigation of and adaption to climate change. Yet, their monitoring and management require significant public resources. remote sensing could facilitate these tasks. Recent hyperspectral satellite programs such as PRISMA have enabled more advanced [...] Read more.
Urban trees support vital ecological functions and help with the mitigation of and adaption to climate change. Yet, their monitoring and management require significant public resources. remote sensing could facilitate these tasks. Recent hyperspectral satellite programs such as PRISMA have enabled more advanced remote sensing applications, such as species classification. However, PRISMA data’s spatial resolution (30 m) could limit its utility in urban areas. Improving hyperspectral data resolution with pansharpening using the PRISMA coregistered panchromatic band (spatial resolution of 5 m) could solve this problem. This study addresses the need to improve hyperspectral data resolution and tests the pansharpening method by classifying exemplative urban tree species in Naples (Italy) using a convolutional neural network and a ground truths dataset, with the aim of comparing results from the original 30 m data to data refined to a 5 m resolution. An evaluation of accuracy metrics shows that pansharpening improves classification quality in dense urban areas with complex topography. In fact, pansharpened data led to significantly higher accuracy for all the examined species. Specifically, the Pinus pinea and Tilia x europaea classes showed an increase of 10% to 20% in their F1 scores. Pansharpening is seen as a practical solution to enhance PRISMA data usability in urban environments. Full article
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>Study area—Region of interest (ROI).</p>
Full article ">Figure 2
<p>This figure shows the principal land cover class present and the altitude of the ROI: (<b>a</b>) using the ISPRA 2021 Land Cover Map; (<b>b</b>) using the DTM of the metropolitan city of Naples.</p>
Full article ">Figure 3
<p>This figure shows the main step used to separate tree canopies from other land covers using the MS Pléiades data: (<b>a</b>) map of vegetated areas with NDVI &gt; 0.2; (<b>b</b>) map of vegetated areas higher than 5 m.</p>
Full article ">Figure 4
<p>This figure shows the two PRISMA HSIs clipped on the ROI in natural colors: (<b>a</b>) HSI1 with 30 m spatial resolution; (<b>b</b>) HSI2 with 5 m spatial resolution.</p>
Full article ">Figure 5
<p>Classification results obtained in all trials (T) using HSI1 and HSI2.</p>
Full article ">
21 pages, 57724 KiB  
Article
MDSCNN: Remote Sensing Image Spatial–Spectral Fusion Method via Multi-Scale Dual-Stream Convolutional Neural Network
by Wenqing Wang, Fei Jia, Yifei Yang, Kunpeng Mu and Han Liu
Remote Sens. 2024, 16(19), 3583; https://doi.org/10.3390/rs16193583 - 26 Sep 2024
Viewed by 467
Abstract
Pansharpening refers to enhancing the spatial resolution of multispectral images through panchromatic images while preserving their spectral features. However, existing traditional methods or deep learning methods always have certain distortions in the spatial or spectral dimensions. This paper proposes a remote sensing spatial–spectral [...] Read more.
Pansharpening refers to enhancing the spatial resolution of multispectral images through panchromatic images while preserving their spectral features. However, existing traditional methods or deep learning methods always have certain distortions in the spatial or spectral dimensions. This paper proposes a remote sensing spatial–spectral fusion method based on a multi-scale dual-stream convolutional neural network, which includes feature extraction, feature fusion, and image reconstruction modules for each scale. In terms of feature fusion, we propose a multi cascade module to better fuse image features. We also design a new loss function aim at enhancing the high degree of consistency between fused images and reference images in terms of spatial details and spectral information. To validate its effectiveness, we conduct thorough experimental analyses on two widely used remote sensing datasets: GeoEye-1 and Ikonos. Compared with the nine leading pansharpening techniques, the proposed method demonstrates superior performance in multiple key evaluation metrics. Full article
Show Figures

Figure 1

Figure 1
<p>The multi-scale dual-stream fusion network framework.</p>
Full article ">Figure 2
<p>The structure of the LCMs. (<b>a</b>) A diagram of the LCM modules; (<b>b</b>) a diagram of the residual module.</p>
Full article ">Figure 3
<p>Decoder module. (<b>a</b>) Overall structure of the decoder; (<b>b</b>) a diagram of the residual module.</p>
Full article ">Figure 4
<p>The fused images of simulated experiments on the GeoEye-1 dataset.</p>
Full article ">Figure 5
<p>The residual images corresponding to the fusion results in <a href="#remotesensing-16-03583-f004" class="html-fig">Figure 4</a>.</p>
Full article ">Figure 6
<p>The fused images of the simulated experiments on the Ikonos dataset.</p>
Full article ">Figure 7
<p>The residual images corresponding to the fusion results in <a href="#remotesensing-16-03583-f006" class="html-fig">Figure 6</a>.</p>
Full article ">Figure 8
<p>The fused images of the real experiments on the Ikonos dataset.</p>
Full article ">Figure 9
<p>The fused images of real experiments on the GeoEye-1 dataset.</p>
Full article ">
20 pages, 8709 KiB  
Article
Automatic Fine Co-Registration of Datasets from Extremely High Resolution Satellite Multispectral Scanners by Means of Injection of Residues of Multivariate Regression
by Luciano Alparone, Alberto Arienzo and Andrea Garzelli
Remote Sens. 2024, 16(19), 3576; https://doi.org/10.3390/rs16193576 - 25 Sep 2024
Viewed by 403
Abstract
This work presents two pre-processing patches to automatically correct the residual local misalignment of datasets acquired by very/extremely high resolution (VHR/EHR) satellite multispectral (MS) scanners, one for, e.g., GeoEye-1 and Pléiades, featuring two separate instruments for MS and panchromatic (Pan) data, the other [...] Read more.
This work presents two pre-processing patches to automatically correct the residual local misalignment of datasets acquired by very/extremely high resolution (VHR/EHR) satellite multispectral (MS) scanners, one for, e.g., GeoEye-1 and Pléiades, featuring two separate instruments for MS and panchromatic (Pan) data, the other for WorldView-2/3 featuring three instruments, two of which are visible and near-infra-red (VNIR) MS scanners. The misalignment arises because the two/three instruments onboard GeoEye-1 / WorldView-2 (four onboard WorldView-3) share the same optics and, thus, cannot have parallel optical axes. Consequently, they image the same swath area from different positions along the orbit. Local height changes (hills, buildings, trees, etc.) originate local shifts among corresponding points in the datasets. The latter would be accurately aligned only if the digital elevation surface model were known with sufficient spatial resolution, which is hardly feasible everywhere because of the extremely high resolution, with Pan pixels of less than 0.5 m. The refined co-registration is achieved by injecting the residue of the multivariate linear regression of each scanner towards lowpass-filtered Pan. Experiments with two and three instruments show that an almost perfect alignment is achieved. MS pansharpening is also shown to greatly benefit from the improved alignment. The proposed alignment procedures are real-time, fully automated, and do not require any additional or ancillary information, but rely uniquely on the unimodality of the MS and Pan sensors. Full article
Show Figures

Figure 1

Figure 1
<p>Laboratory MTF of a spectral channel of the Pléiades instrument.</p>
Full article ">Figure 2
<p>Spectral responsivity functions of: (<b>left</b>) GeoEye-1, bands: blue, Pan, green, red, NIR. Left to right; (<b>right</b>) WorldView-2, bands coastal, blue, Pan, green, yellow, red edge, NIR1, NIR2, with one MS scanner (MS1) taking the same four bands as GeoEye-1; another MS scanner (MS2) capturing coastal, yellow, red edge and NIR2.</p>
Full article ">Figure 3
<p>Acquisition geometries of GeoEye-1 (<b>left</b>) and WorldView-2/3 (<b>right</b>) MS scanning systems. In the former case, there are two separate instruments, one for the four MS bands and another for Pan, whose optical axes are not perfectly parallel. The acquisition of the same scan line orthogonal to the ground track of the platform occurs on different instants; hence, from different points along the orbit, the acquisition angles from the orbit are deliberately exaggerated. In the latter case, the plot is simplified: the optical axis of the one out of three instruments that performs the acquisition from the respective positions along the orbit is shown with a solid dark line.</p>
Full article ">Figure 4
<p>512 × 512 portion of test GeoEye-1 image at 0.5 m: (<b>a</b>) Pan; (<b>b</b>) interpolated MS B3-B2-B1 (true color).</p>
Full article ">Figure 5
<p>Residue of multivariate regression to lowpass-filtered Pan of: (<b>a</b>) resampled MS bands, <math display="inline"><semantics> <mrow> <msup> <mi>R</mi> <mn>2</mn> </msup> <mo>=</mo> <mn>0.93914</mn> </mrow> </semantics></math>; (<b>b</b>) corrected MS bands, <math display="inline"><semantics> <mrow> <msup> <mi>R</mi> <mn>2</mn> </msup> <mo>=</mo> <mn>0.99810</mn> </mrow> </semantics></math>.</p>
Full article ">Figure 6
<p>Visual results of fusion without correction for MS-to-Pan alignment: (<b>a</b>) BT-H without alignment; (<b>b</b>) GSA without alignment; (<b>c</b>) AWLP-H without alignment; (<b>d</b>) MTF-GLP-FS without alignment.</p>
Full article ">Figure 7
<p>Visual results of fusion with correction for MS-to-Pan alignment: (<b>a</b>) BT-H with alignment; (<b>b</b>) GSA with alignment; (<b>c</b>) AWLP-H with alignment; (<b>d</b>) MTF-GLP-FS with alignment.</p>
Full article ">Figure 8
<p>512 × 512 portion of test WorldView-2 image at 0.5 m: (<b>a</b>) Pan; (<b>b</b>) interpolated B5-B3-B2 (true color); (<b>c</b>) interpolated B6-B4-B1 (false color). Each image displayed is taken by a unique instrument without combining the spectral bands of different scanners.</p>
Full article ">Figure 9
<p>Residue of multivariate regression to lowpass-filtered Pan of: (<b>a</b>) the four bands of MS1, <math display="inline"><semantics> <mrow> <msup> <mi>R</mi> <mn>2</mn> </msup> <mo>=</mo> <mn>0.97954</mn> </mrow> </semantics></math>; (<b>b</b>) the four bands of MS2, <math display="inline"><semantics> <mrow> <msup> <mi>R</mi> <mn>2</mn> </msup> <mo>=</mo> <mn>0.97846</mn> </mrow> </semantics></math>; (<b>c</b>) the eight bands together (MS1 + MS2), <math display="inline"><semantics> <mrow> <msup> <mi>R</mi> <mn>2</mn> </msup> <mo>=</mo> <mn>0.984490</mn> </mrow> </semantics></math>; (<b>d</b>) the corrected bands of MS1, <math display="inline"><semantics> <mrow> <msup> <mi>R</mi> <mn>2</mn> </msup> <mo>=</mo> <mn>0.99993</mn> </mrow> </semantics></math>; (<b>e</b>) the corrected bands of MS2, <math display="inline"><semantics> <mrow> <msup> <mi>R</mi> <mn>2</mn> </msup> <mo>=</mo> <mn>0.99968</mn> </mrow> </semantics></math>; (<b>f</b>) MS1+MS2 after correction, <math display="inline"><semantics> <mrow> <msup> <mi>R</mi> <mn>2</mn> </msup> <mo>=</mo> <mn>0.99999</mn> </mrow> </semantics></math>. All residue maps have been linearly stretched by the same factor for displaying convenience.</p>
Full article ">Figure 10
<p>Fusion results of a CS method (BT-H): (<b>a</b>) B5-B3-B2 composition without alignment; (<b>b</b>) B5-B3-B2 with alignment; (<b>c</b>) B6-B3-B1 without alignment; (<b>d</b>) B6-B3-B1 with alignment.</p>
Full article ">Figure 11
<p>Visual results of BT-H fusion with and without correction for MS-to-Pan alignment: (<b>a</b>) 5-3-2 composition (R-G-B true color captured by MS1) without alignment; (<b>b</b>) 6-4-1 composition (RE-Y-C false color captured by MS2) without alignment; (<b>c</b>) 6-3-1 composition (RE-G-C false color captured by MS1 and MS2 together) without alignment; (<b>d</b>) 5-3-2 composition (R-G-B true color captured by MS1) with alignment; (<b>e</b>) 6-4-1 composition (RE-Y-C false color captured by MS2) with alignment; (<b>f</b>) 6-3-1 composition (RE-G-C false color captured by MS1 and MS2 together) with alignment.</p>
Full article ">Figure 12
<p>Visual results of GSA fusion with and without correction for MS-to-Pan alignment: (<b>a</b>) 5-3-2 composition (R-G-B true color captured by MS1) without alignment; (<b>b</b>) 6-4-1 composition (RE-Y-C false color captured by MS2) without alignment; (<b>c</b>) 6-3-1 composition (RE-G-C false color captured by MS1 and MS2 together) without alignment; (<b>d</b>) 5-3-2 composition (R-G-B true color captured by MS1) with alignment; (<b>e</b>) 6-4-1 composition (RE-Y-C false color captured by MS2) with alignment; (<b>f</b>) 6-3-1 composition (RE-G-C false color captured by MS1 and MS2 together) with alignment.</p>
Full article ">Figure 13
<p>Visual results of AWLP-H fusion with and without correction for MS-to-Pan alignment: (<b>a</b>) 5-3-2 composition (R-G-B true color captured by MS1) without alignment; (<b>b</b>) 6-4-1 composition (RE-Y-C false color captured by MS2) without alignment; (<b>c</b>) 6-3-1 composition (RE-G-C false color captured by MS1 and MS2 together) without alignment; (<b>d</b>) 5-3-2 composition (R-G-B true color captured by MS1) with alignment; (<b>e</b>) 6-4-1 composition (RE-Y-C false color captured by MS2) with alignment; (<b>f</b>) 6-3-1 composition (RE-G-C false color captured by MS1 and MS2 together) with alignment.</p>
Full article ">Figure 14
<p>Visual results of MTF-GLP-FS fusion with and without correction for MS-to-Pan alignment: (<b>a</b>) 5-3-2 composition (R-G-B true color captured by MS1) without alignment; (<b>b</b>) 6-4-1 composition (RE-Y-C false color captured by MS2) without alignment; (<b>c</b>) 6-3-1 composition (RE-G-C false color captured by MS1 and MS2 together) without alignment; (<b>d</b>) 5-3-2 composition (R-G-B true color captured by MS1) with alignment; (<b>e</b>) 6-4-1 composition (RE-Y-C false color captured by MS2) with alignment; (<b>f</b>) 6-3-1 composition (RE-G-C false color captured by MS1 and MS2 together) with alignment.</p>
Full article ">
23 pages, 30982 KiB  
Article
A Multi-Stage Progressive Pansharpening Network Based on Detail Injection with Redundancy Reduction
by Xincan Wen, Hongbing Ma and Liangliang Li
Sensors 2024, 24(18), 6039; https://doi.org/10.3390/s24186039 - 18 Sep 2024
Viewed by 376
Abstract
In the field of remote sensing image processing, pansharpening technology stands as a critical advancement. This technology aims to enhance multispectral images that possess low resolution by integrating them with high-spatial-resolution panchromatic images, ultimately producing multispectral images with high resolution that are abundant [...] Read more.
In the field of remote sensing image processing, pansharpening technology stands as a critical advancement. This technology aims to enhance multispectral images that possess low resolution by integrating them with high-spatial-resolution panchromatic images, ultimately producing multispectral images with high resolution that are abundant in both spatial and spectral details. Thus, there remains potential for improving the quality of both the spectral and spatial domains of the fused images based on deep-learning-based pansharpening methods. This work proposes a new method for the task of pansharpening: the Multi-Stage Progressive Pansharpening Network with Detail Injection with Redundancy Reduction Mechanism (MSPPN-DIRRM). This network is divided into three levels, each of which is optimized for the extraction of spectral and spatial data at different scales. Particular spectral feature and spatial detail extraction modules are used at each stage. Moreover, a new image reconstruction module named the DRRM is introduced in this work; it eliminates both spatial and channel redundancy and improves the fusion quality. The effectiveness of the proposed model is further supported by experimental results using both simulated data and real data from the QuickBird, GaoFen1, and WorldView2 satellites; these results show that the proposed model outperforms deep-learning-based methods in both visual and quantitative assessments. Among various evaluation metrics, performance improves by 0.92–18.7% compared to the latest methods. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

Figure 1
<p>The architecture of the proposed method.</p>
Full article ">Figure 2
<p>The architecture of the CFEM.</p>
Full article ">Figure 3
<p>(<b>a</b>) The architecture of the overall SDEM. (<b>b</b>) The architecture of spatial attention.</p>
Full article ">Figure 4
<p>The architecture of the overall DRRM.</p>
Full article ">Figure 5
<p>The architecture of the SRRM.</p>
Full article ">Figure 6
<p>The architecture of the CRRM.</p>
Full article ">Figure 7
<p>The fusion results of various methods on the QB simulation dataset.</p>
Full article ">Figure 8
<p>The absolute error maps between the fusion results of all methods and the reference image on the QB simulation dataset.</p>
Full article ">Figure 9
<p>The fusion results of various methods on the GaoFen1 simulation dataset.</p>
Full article ">Figure 10
<p>The absolute error maps between the fusion results of all methods and the reference image on the GaoFen1 simulation dataset.</p>
Full article ">Figure 11
<p>The fusion results of various methods on the WV2 simulation dataset.</p>
Full article ">Figure 12
<p>The absolute error maps between the fusion results of all methods and the reference image on the WV2 simulation dataset.</p>
Full article ">Figure 13
<p>The fusion results of various methods on the QB real dataset.</p>
Full article ">Figure 14
<p>The fusion results of various methods on the GaoFen1 real dataset.</p>
Full article ">Figure 15
<p>The fusion results of various methods on the WV2 real dataset.</p>
Full article ">Figure 16
<p>The fusion results after the removal of certain modules. (<b>a</b>) LRMS image, (<b>b</b>) PAN image, (<b>c</b>) reference image, (<b>d</b>) w/o DRRM, (<b>e</b>) w/o CFEM, (<b>f</b>) w/o SDEM, (<b>g</b>) w/o (SDEM + DRRM), (<b>h</b>) w/o (CFEM + DRRM), (<b>i</b>) w/o (CFEM + SDEM), (<b>j</b>) none_modules, and (<b>k</b>) ours.</p>
Full article ">Figure 17
<p>Residual plots comparing the outcomes of the experiments with different module omissions against the reference image. (<b>a</b>) Reference image, (<b>b</b>) w/o DRRM, (<b>c</b>) w/o CFEM, (<b>d</b>) w/o SDEM, (<b>e</b>) w/o (SDEM + DRRM), (<b>f</b>) w/o (CFEM + DRRM), (<b>g</b>) w/o (CFEM + SDEM), (<b>h</b>) none_modules, and (<b>i</b>) ours.</p>
Full article ">Figure 18
<p>The output results of the various network structures along with their residual plots relative to the reference image. (<b>a</b>) Reference image, (<b>b</b>) concatenation, (<b>c</b>) direct injection, (<b>d</b>) one-stage, (<b>e</b>) two-stage, and (<b>f</b>) Ours.</p>
Full article ">
17 pages, 5435 KiB  
Article
HyperKon: A Self-Supervised Contrastive Network for Hyperspectral Image Analysis
by Daniel La’ah Ayuba, Jean-Yves Guillemaut, Belen Marti-Cardona and Oscar Mendez
Remote Sens. 2024, 16(18), 3399; https://doi.org/10.3390/rs16183399 - 12 Sep 2024
Viewed by 569
Abstract
The use of a pretrained image classification model (trained on cats and dogs, for example) as a perceptual loss function for hyperspectral super-resolution and pansharpening tasks is surprisingly effective. However, RGB-based networks do not take full advantage of the spectral information in hyperspectral [...] Read more.
The use of a pretrained image classification model (trained on cats and dogs, for example) as a perceptual loss function for hyperspectral super-resolution and pansharpening tasks is surprisingly effective. However, RGB-based networks do not take full advantage of the spectral information in hyperspectral data. This inspired the creation of HyperKon, a dedicated hyperspectral Convolutional Neural Network backbone built with self-supervised contrastive representation learning. HyperKon uniquely leverages the high spectral continuity, range, and resolution of hyperspectral data through a spectral attention mechanism. We also perform a thorough ablation study on different kinds of layers, showing their performance in understanding hyperspectral layers. Notably, HyperKon achieves a remarkable 98% Top-1 retrieval accuracy and surpasses traditional RGB-trained backbones in both pansharpening and image classification tasks. These results highlight the potential of hyperspectral-native backbones and herald a paradigm shift in hyperspectral image analysis. Full article
(This article belongs to the Special Issue Advances in Hyperspectral Remote Sensing Image Processing)
Show Figures

Figure 1

Figure 1
<p>General Overview of HyperKon System Architecture.</p>
Full article ">Figure 2
<p>Illustration of HSI contrastive sampling.</p>
Full article ">Figure 3
<p>Conceptual comparison of HSI vs. RGB perceptual loss.</p>
Full article ">Figure 4
<p>Top-1 HSI retrieval accuracy achieved by various versions of the HyperKon model during the pretraining phase. The performance of each version is presented as a bar in the chart, illustrating how the integration of different components, such as 3D convolutions, DSC, the CBAM, and the SEB, affected the model’s accuracy. The chart underscores the superior performance of the SEB.</p>
Full article ">Figure 5
<p>Top-1 HSI retrieval accuracy for the 3D Conv, SEB, and CBAM models following dimensionality reduction using PCA. The graph indicates a superior performance by the 3D convolution model.</p>
Full article ">Figure 6
<p>Top-1 HSI retrieval accuracy for the 3D Conv, SEB, and CBAM models when manual band selection was employed. It shows an initial advantage for 3D Conv, but over time, the SEB and CBAM models catch up to similar levels of performance.</p>
Full article ">Figure 7
<p>Visual results generated by different pansharpening algorithms (HyperPNN [<a href="#B49-remotesensing-16-03399" class="html-bibr">49</a>], DARN [<a href="#B45-remotesensing-16-03399" class="html-bibr">45</a>], GPPNN [<a href="#B50-remotesensing-16-03399" class="html-bibr">50</a>], HyperTransformer [<a href="#B51-remotesensing-16-03399" class="html-bibr">51</a>], HyperKon (ours), and ground truth) for Pavia Center [<a href="#B42-remotesensing-16-03399" class="html-bibr">42</a>], Botswana [<a href="#B43-remotesensing-16-03399" class="html-bibr">43</a>], Chikusei [<a href="#B44-remotesensing-16-03399" class="html-bibr">44</a>], and EnMAP [<a href="#B26-remotesensing-16-03399" class="html-bibr">26</a>] datasets. MAE denotes the (normalized) Mean Absolute Error across all spectral bands.</p>
Full article ">Figure 8
<p>HyperKon image classification visualization for Indian Pines, Pavia University, and Salinas datasets. (<b>a</b>) Predicted classification map generated by HyperKon, (<b>b</b>) Predicted classification map with masked regions (showing only labeled areas), (<b>c</b>) predicted accuracy map: green for correct predictions, red for incorrect predictions, and black for unlabeled areas, (<b>d</b>) ground-truth classification map, and (<b>e</b>) original RGB image.</p>
Full article ">Figure 9
<p>Zoom-out predicted accuracy map for Indian Pines: green for correct predictions, red for incorrect predictions, and black for unlabeled areas.</p>
Full article ">
18 pages, 2642 KiB  
Article
An Unsupervised CNN-Based Pansharpening Framework with Spectral-Spatial Fidelity Balance
by Matteo Ciotola, Giuseppe Guarino and Giuseppe Scarpa
Remote Sens. 2024, 16(16), 3014; https://doi.org/10.3390/rs16163014 - 16 Aug 2024
Viewed by 571
Abstract
In recent years, deep learning techniques for pansharpening multiresolution images have gained increasing interest. Due to the lack of ground truth data, most deep learning solutions rely on synthetic reduced-resolution data for supervised training. This approach has limitations due to the statistical mismatch [...] Read more.
In recent years, deep learning techniques for pansharpening multiresolution images have gained increasing interest. Due to the lack of ground truth data, most deep learning solutions rely on synthetic reduced-resolution data for supervised training. This approach has limitations due to the statistical mismatch between real full-resolution and synthetic reduced-resolution data, which affects the models’ generalization capacity. Consequently, there has been a shift towards unsupervised learning frameworks for pansharpening deep learning-based techniques. Unsupervised schemes require defining sophisticated loss functions with at least two components: one for spectral quality, ensuring consistency between the pansharpened image and the input multispectral component, and another for spatial quality, ensuring consistency between the output and the panchromatic input. Despite promising results, there has been limited investigation into the interaction and balance of these loss terms to ensure stability and accuracy. This work explores how unsupervised spatial and spectral consistency losses can be reliably combined preserving the outocome quality. By examining these interactions, we propose a general rule for balancing the two loss components to enhance the stability and performance of unsupervised pansharpening models. Experiments on three state-of-the-art algorithms using WorldView-3 images demonstrate that methods trained with the proposed framework achieve good performance in terms of visual quality and numerical indexes. Full article
(This article belongs to the Special Issue Weakly Supervised Deep Learning in Exploiting Remote Sensing Big Data)
Show Figures

Figure 1

Figure 1
<p>A-PNN model for Z-PNN [<a href="#B46-remotesensing-16-03014" class="html-bibr">46</a>].</p>
Full article ">Figure 2
<p>PanNet architecture [<a href="#B11-remotesensing-16-03014" class="html-bibr">11</a>].</p>
Full article ">Figure 3
<p>BDPN model [<a href="#B37-remotesensing-16-03014" class="html-bibr">37</a>].</p>
Full article ">Figure 4
<p>Spectral loss terms after 200 iterations for different <math display="inline"><semantics> <mfenced separators="" open="(" close=")"> <mi>α</mi> <mo>,</mo> <mi>β</mi> </mfenced> </semantics></math> configurations.</p>
Full article ">Figure 5
<p>Spatial loss terms after 200 iterations for different <math display="inline"><semantics> <mfenced separators="" open="(" close=")"> <mi>α</mi> <mo>,</mo> <mi>β</mi> </mfenced> </semantics></math> configurations.</p>
Full article ">Figure 6
<p>Sample results taken from WV3 Adelaide dataset: input MS and PAN followed by related pansharpenings obtained by Z-PNN through target adaptation for 500 epochs with different <math display="inline"><semantics> <mrow> <mo>(</mo> <mi>α</mi> <mo>,</mo> <mi>β</mi> <mo>)</mo> </mrow> </semantics></math> settings.</p>
Full article ">Figure 7
<p>Sample results taken from WV3 Adelaide dataset: input MS and PAN followed by related pansharpenings obtained by PanNet through target adaptation for 500 epochs with different <math display="inline"><semantics> <mrow> <mo>(</mo> <mi>α</mi> <mo>,</mo> <mi>β</mi> <mo>)</mo> </mrow> </semantics></math> settings.</p>
Full article ">Figure 8
<p>Sample results taken from WV3 Adelaide dataset: input MS and PAN followed by related pansharpenings obtained by BDPN through target adaptation for 500 epochs with different <math display="inline"><semantics> <mrow> <mo>(</mo> <mi>α</mi> <mo>,</mo> <mi>β</mi> <mo>)</mo> </mrow> </semantics></math> settings.</p>
Full article ">Figure 9
<p>Full-resolution accuracy indexes for Z-PNN, with <math display="inline"><semantics> <mrow> <mi>α</mi> <mo>=</mo> <mn>0.03</mn> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <mi>β</mi> <mo>∈</mo> <mfenced separators="" open="[" close="]"> <mn>3</mn> <mo>·</mo> <msup> <mn>10</mn> <mrow> <mo>−</mo> <mn>5</mn> </mrow> </msup> <mo>,</mo> <mn>3</mn> <mo>·</mo> <msup> <mn>10</mn> <mrow> <mo>−</mo> <mn>4</mn> </mrow> </msup> <mo>,</mo> <mo>…</mo> <mo>,</mo> <mn>3</mn> <mo>·</mo> <msup> <mn>10</mn> <mn>3</mn> </msup> </mfenced> </mrow> </semantics></math>.</p>
Full article ">Figure 10
<p>Full-resolution accuracy indexes for PanNet, with <math display="inline"><semantics> <mrow> <mi>α</mi> <mo>=</mo> <mn>0.03</mn> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <mi>β</mi> <mo>∈</mo> <mfenced separators="" open="[" close="]"> <mn>3</mn> <mo>·</mo> <msup> <mn>10</mn> <mrow> <mo>−</mo> <mn>5</mn> </mrow> </msup> <mo>,</mo> <mn>3</mn> <mo>·</mo> <msup> <mn>10</mn> <mrow> <mo>−</mo> <mn>4</mn> </mrow> </msup> <mo>,</mo> <mo>…</mo> <mo>,</mo> <mn>3</mn> <mo>·</mo> <msup> <mn>10</mn> <mn>3</mn> </msup> </mfenced> </mrow> </semantics></math>.</p>
Full article ">Figure 11
<p>Full-resolution accuracy indexes for BDPN, with <math display="inline"><semantics> <mrow> <mi>α</mi> <mo>=</mo> <mn>0.03</mn> </mrow> </semantics></math> and <math display="inline"><semantics> <mrow> <mi>β</mi> <mo>∈</mo> <mfenced separators="" open="[" close="]"> <mn>3</mn> <mo>·</mo> <msup> <mn>10</mn> <mrow> <mo>−</mo> <mn>5</mn> </mrow> </msup> <mo>,</mo> <mn>3</mn> <mo>·</mo> <msup> <mn>10</mn> <mrow> <mo>−</mo> <mn>4</mn> </mrow> </msup> <mo>,</mo> <mo>…</mo> <mo>,</mo> <mn>3</mn> <mo>·</mo> <msup> <mn>10</mn> <mn>3</mn> </msup> </mfenced> </mrow> </semantics></math>.</p>
Full article ">
26 pages, 6739 KiB  
Article
Pansharpening Based on Multimodal Texture Correction and Adaptive Edge Detail Fusion
by Danfeng Liu, Enyuan Wang, Liguo Wang, Jón Atli Benediktsson, Jianyu Wang and Lei Deng
Remote Sens. 2024, 16(16), 2941; https://doi.org/10.3390/rs16162941 - 11 Aug 2024
Viewed by 669
Abstract
Pansharpening refers to the process of fusing multispectral (MS) images with panchromatic (PAN) images to obtain high-resolution multispectral (HRMS) images. However, due to the low correlation and similarity between MS and PAN images, as well as inaccuracies in spatial information injection, HRMS images [...] Read more.
Pansharpening refers to the process of fusing multispectral (MS) images with panchromatic (PAN) images to obtain high-resolution multispectral (HRMS) images. However, due to the low correlation and similarity between MS and PAN images, as well as inaccuracies in spatial information injection, HRMS images often suffer from significant spectral and spatial distortions. To address these issues, a pansharpening method based on multimodal texture correction and adaptive edge detail fusion is proposed in this paper. To obtain a texture-corrected (TC) image that is highly correlated and similar to the MS image, the target-adaptive CNN-based pansharpening (A-PNN) method is introduced. By constructing a multimodal texture correction model, intensity, gradient, and A-PNN-based deep plug-and-play correction constraints are established between the TC and source images. Additionally, an adaptive degradation filter algorithm is proposed to ensure the accuracy of these constraints. Since the TC image obtained can effectively replace the PAN image and considering that the MS image contains valuable spatial information, an adaptive edge detail fusion algorithm is also proposed. This algorithm adaptively extracts detailed information from the TC and MS images to apply edge protection. Given the limited spatial information in the MS image, its spatial information is proportionally enhanced before the adaptive fusion. The fused spatial information is then injected into the upsampled multispectral (UPMS) image to produce the final HRMS image. Extensive experimental results demonstrated that compared with other methods, the proposed algorithm achieved superior results in terms of both subjective visual effects and objective evaluation metrics. Full article
Show Figures

Figure 1

Figure 1
<p>The proposed model framework diagram.</p>
Full article ">Figure 2
<p>Iterative convergence results from the WorldView-3 dataset.</p>
Full article ">Figure 3
<p>Subjective evaluation fusion results of the RS images in the QuickBird dataset.</p>
Full article ">Figure 4
<p>Subjective evaluation fusion results of the RS images in the WorldView-2 dataset.</p>
Full article ">Figure 5
<p>Subjective evaluation fusion results of the RS images in the WorldView-3 dataset.</p>
Full article ">Figure 6
<p>Subjective evaluation fusion results of the FS images in the GaoFen-2 dataset.</p>
Full article ">Figure 7
<p>Subjective evaluation fusion results of the FS images in the WorldView-2 dataset.</p>
Full article ">Figure 8
<p>Parameter settings for four different datasets in RS and FS experiments.</p>
Full article ">Figure 9
<p>Subjective evaluation fusion results of different ablation combination models from the WorldView-3 dataset.</p>
Full article ">
22 pages, 7835 KiB  
Article
Towards Robust Pansharpening: A Large-Scale High-Resolution Multi-Scene Dataset and Novel Approach
by Shiying Wang, Xuechao Zou, Kai Li, Junliang Xing, Tengfei Cao and Pin Tao
Remote Sens. 2024, 16(16), 2899; https://doi.org/10.3390/rs16162899 - 8 Aug 2024
Cited by 1 | Viewed by 993
Abstract
Pansharpening, a pivotal task in remote sensing, involves integrating low-resolution multispectral images with high-resolution panchromatic images to synthesize an image that is both high-resolution and retains multispectral information. These pansharpened images enhance precision in land cover classification, change detection, and environmental monitoring within [...] Read more.
Pansharpening, a pivotal task in remote sensing, involves integrating low-resolution multispectral images with high-resolution panchromatic images to synthesize an image that is both high-resolution and retains multispectral information. These pansharpened images enhance precision in land cover classification, change detection, and environmental monitoring within remote sensing data analysis. While deep learning techniques have shown significant success in pansharpening, existing methods often face limitations in their evaluation, focusing on restricted satellite data sources, single scene types, and low-resolution images. This paper addresses this gap by introducing PanBench, a high-resolution multi-scene dataset containing all mainstream satellites and comprising 5898 pairs of samples. Each pair includes a four-channel (RGB + near-infrared) multispectral image of 256 × 256 pixels and a mono-channel panchromatic image of 1024 × 1024 pixels. To avoid irreversible loss of spectral information and achieve a high-fidelity synthesis, we propose a Cascaded Multiscale Fusion Network (CMFNet) for pansharpening. Multispectral images are progressively upsampled while panchromatic images are downsampled. Corresponding multispectral features and panchromatic features at the same scale are then fused in a cascaded manner to obtain more robust features. Extensive experiments validate the effectiveness of CMFNet. Full article
Show Figures

Figure 1

Figure 1
<p>All mainstream satellites for pansharpening are supported by our PanBench dataset, along with their respective sample quantities.</p>
Full article ">Figure 2
<p>Data preprocessing flow chart.</p>
Full article ">Figure 3
<p>Number of the six scenes of land cover classification.</p>
Full article ">Figure 4
<p>The overall framework of our proposed CMFNet, where <math display="inline"><semantics> <msub> <mi mathvariant="bold">F</mi> <mi>o</mi> </msub> </semantics></math> represent the output of the autoencoder.</p>
Full article ">Figure 5
<p>The detailed structure of the three components in CMFNet. The upward and downward arrows indicate upsampling and downsampling, respectively, while the plus sign denotes summation. The term “block” can be substituted by any module, such as ResNet [<a href="#B42-remotesensing-16-02899" class="html-bibr">42</a>], NAFNet [<a href="#B51-remotesensing-16-02899" class="html-bibr">51</a>], ConvNeXt [<a href="#B52-remotesensing-16-02899" class="html-bibr">52</a>] block, etc. The autoencoder integrates spectral and spatial information by incorporating identical-scale features from injecting two encoders.</p>
Full article ">Figure 6
<p>Visual comparison of GF1 and GF2 images from different pansharpening methods. MAE represents the mean absolute error of each spectral band.</p>
Full article ">Figure 7
<p>Visual comparison of QB and IN images from different pansharpening methods. MAE represents the mean absolute error of each spectral band.</p>
Full article ">Figure 8
<p>Visual results generated by different pansharpening methods for water in the scene category (from GF6). The first line of images is displayed in RGB combination; the second line is the MAE difference between the output pansharpening result and the ground truth.</p>
Full article ">Figure 9
<p>Visual results generated by different pansharpening methods for urban in the scene category (from WV4). The first line of images is displayed in RGB combination; the second line is the MAE difference between the output pansharpening result and the ground truth.</p>
Full article ">Figure 10
<p>Visual results generated by different pansharpening methods for ice/snow in the scene category (from LC7). The first line of images is displayed in RGB combination; the second line is the MAE difference between the output pansharpening result and the ground truth.</p>
Full article ">Figure 11
<p>Visual results generated by different pansharpening methods for crops in the scene category (from WV3). The first line of images is displayed in RGB combination; the second line is the MAE difference between the output pansharpening result and the ground truth.</p>
Full article ">Figure 12
<p>Visual results generated by different pansharpening methods for vegetation in the scene category (from WV2). The first line of images is displayed in RGB combination; the second line is the MAE difference between the output pansharpening result and the ground truth.</p>
Full article ">Figure 13
<p>Visual results generated by different pansharpening methods for barren in the scene category (from LC8). The first line of images is displayed in RGB combination; the second line is the MAE difference between the output pansharpening result and the ground truth.</p>
Full article ">Figure 14
<p>Example results for full-resolution experiments. Displayed in RGB combination.</p>
Full article ">
21 pages, 5849 KiB  
Article
MMCMOO: A Novel Multispectral Pansharpening Method
by Yingxia Chen and Yingying Xu
Mathematics 2024, 12(14), 2255; https://doi.org/10.3390/math12142255 - 19 Jul 2024
Viewed by 414
Abstract
From the perspective of optimization, most of the current mainstream remote sensing data fusion methods are based on traditional mathematical optimization or single objective optimization. The former requires manual parameter tuning and easily falls into local optimum. Although the latter can overcome the [...] Read more.
From the perspective of optimization, most of the current mainstream remote sensing data fusion methods are based on traditional mathematical optimization or single objective optimization. The former requires manual parameter tuning and easily falls into local optimum. Although the latter can overcome the shortcomings of traditional methods, the single optimization objective makes it unable to combine the advantages of multiple models, which may lead to distortion of the fused image. To address the problems of missing multi-model combination and parameters needing to be set manually in the existing methods, a pansharpening method based on multi-model collaboration and multi-objective optimization is proposed, called MMCMOO. In the proposed new method, the multi-spectral image fusion problem is transformed into a multi-objective optimization problem. Different evolutionary strategies are used to design a variety of population generation mechanisms, and a non-dominated sorting genetic algorithm (NSGA-II) is used to optimize the two proposed target models, so as to obtain the best pansharpening quality. The experimental results show that the proposed method is superior to the traditional methods and single objective methods in terms of visual comparison and quantitative analysis on our datasets. Full article
Show Figures

Figure 1

Figure 1
<p>The technical route of MMCMOO.</p>
Full article ">Figure 2
<p>The chromosome representation.</p>
Full article ">Figure 3
<p>Population generation strategy.</p>
Full article ">Figure 4
<p>Pareto optimal solution selection strategy.</p>
Full article ">Figure 5
<p>One-step local search operator.</p>
Full article ">Figure 6
<p>Pareto Frontier and Crowding Distance.</p>
Full article ">Figure 7
<p>Flow chart of MMCMOO optimization.</p>
Full article ">Figure 8
<p>Pansharpening results (source: Google Earth). (<b>a</b>) Original MS; (<b>b</b>) PAN; (<b>c</b>) the AIHS method; (<b>d</b>) the wavelet method; (<b>e</b>) the PCA method; (<b>f</b>) the EIHS method; (<b>g</b>) the PAIHS method; (<b>h</b>) the MMCMOO method. PAN image resolution is 256 × 256.</p>
Full article ">Figure 9
<p>Pansharpening results (source: Pléiades data). (<b>a</b>) Original MS; (<b>b</b>) PAN; (<b>c</b>) the AIHS method; (<b>d</b>) the wavelet method; (<b>e</b>) the PCA method; (<b>f</b>) the EIHS method; (<b>g</b>) the PAIHS method; (<b>h</b>) the MMCMOO method. PAN image resolution is 256 × 256.</p>
Full article ">Figure 10
<p>Pansharpening results (source: Google Earth). (<b>a</b>) Original MS; (<b>b</b>) PAN; (<b>c</b>) the AIHS method; (<b>d</b>) the wavelet method; (<b>e</b>) the PCA method; (<b>f</b>) the EIHS method; (<b>g</b>) the PAIHS method; (<b>h</b>) the MMCMOO method. PAN image resolution is 256 × 256.</p>
Full article ">Figure 11
<p>Comparison of pansharpening results (source: Google Earth). (<b>a</b>) Original MS; (<b>b</b>) PAN; (<b>c</b>) AIHS; (<b>d</b>) Wavelet; (<b>e</b>) PCA; (<b>f</b>) EIHS; (<b>g</b>) PAIHS; (<b>h</b>) MMCMOO. PAN image resolution is 256 × 256.</p>
Full article ">Figure 12
<p>Comparison of pansharpening results (source: QuickBird). (<b>a</b>) Original MS; (<b>b</b>) PAN; (<b>c</b>) the AIHS method; (<b>d</b>) the wavelet method; (<b>e</b>) the PCA method; (<b>f</b>) the EIHS method; (<b>g</b>) the PAIHS method; (<b>h</b>) the MMCMOO method. PAN image resolution is 256 × 256.</p>
Full article ">Figure 12 Cont.
<p>Comparison of pansharpening results (source: QuickBird). (<b>a</b>) Original MS; (<b>b</b>) PAN; (<b>c</b>) the AIHS method; (<b>d</b>) the wavelet method; (<b>e</b>) the PCA method; (<b>f</b>) the EIHS method; (<b>g</b>) the PAIHS method; (<b>h</b>) the MMCMOO method. PAN image resolution is 256 × 256.</p>
Full article ">Figure 13
<p>Convergence analysis on different quantitative evaluation metrics.</p>
Full article ">Figure 14
<p>Each row from top to bottom corresponds to optimization Model Z2, Model Z1, and the full model, respectively. From left to right, each column corresponds to the red, green and blue bands, respectively.</p>
Full article ">Figure 15
<p>Relationship diagram between evaluation index and model selection.</p>
Full article ">
16 pages, 4099 KiB  
Article
Multi-Frequency Spectral–Spatial Interactive Enhancement Fusion Network for Pan-Sharpening
by Yunxuan Tang, Huaguang Li, Guangxu Xie, Peng Liu and Tong Li
Electronics 2024, 13(14), 2802; https://doi.org/10.3390/electronics13142802 - 16 Jul 2024
Viewed by 506
Abstract
The objective of pan-sharpening is to effectively fuse high-resolution panchromatic (PAN) images with limited spectral information and low-resolution multispectral (LR-MS) images, thereby generating a fused image with a high spatial resolution and rich spectral information. However, current fusion techniques face significant challenges, including [...] Read more.
The objective of pan-sharpening is to effectively fuse high-resolution panchromatic (PAN) images with limited spectral information and low-resolution multispectral (LR-MS) images, thereby generating a fused image with a high spatial resolution and rich spectral information. However, current fusion techniques face significant challenges, including insufficient edge detail, spectral distortion, increased noise, and limited robustness. To address these challenges, we propose a multi-frequency spectral–spatial interaction enhancement network (MFSINet) that comprises the spectral–spatial interactive fusion (SSIF) and multi-frequency feature enhancement (MFFE) subnetworks. The SSIF enhances both spatial and spectral fusion features by optimizing the characteristics of each spectral band through band-aware processing. The MFFE employs a variant of wavelet transform to perform multiresolution analyses on remote sensing scenes, enhancing the spatial resolution, spectral fidelity, and the texture and structural features of the fused images by optimizing directional and spatial properties. Moreover, qualitative analysis and quantitative comparative experiments using the IKONOS and WorldView-2 datasets indicate that this method significantly improves the fidelity and accuracy of the fused images. Full article
(This article belongs to the Topic Computational Intelligence in Remote Sensing: 2nd Edition)
Show Figures

Figure 1

Figure 1
<p>Framework of the proposed MFSINet.</p>
Full article ">Figure 2
<p>Framework of the proposed SSIF.</p>
Full article ">Figure 3
<p>Framework of the proposed MFFE.</p>
Full article ">Figure 4
<p>Results images of the nine methods and GT on the IKONOS simulated dataset, as well as images of the absolute error.</p>
Full article ">Figure 5
<p>Resulting images of the nine methods and GT on the WV-2 simulated dataset, as well as images of the absolute error.</p>
Full article ">Figure 6
<p>Resulting images of the nine methods on the IKONOS real dataset. The lower part indicates the magnified details of the fused results (red and blue boxes).</p>
Full article ">Figure 7
<p>Resulting images of the nine methods on the WV-2 real dataset. The lower part indicates the magnified details of the fused results (red and blue boxes).</p>
Full article ">Figure 8
<p>Resulting images of different types of ablation experiments on the IKONOS (<b>top</b>) and WV-2 (<b>bottom</b>) simulated datasets, along with the absolute error images.</p>
Full article ">
27 pages, 10814 KiB  
Article
UPGAN: An Unsupervised Generative Adversarial Network Based on U-Shaped Structure for Pansharpening
by Xin Jin, Yuting Feng, Qian Jiang, Shengfa Miao, Xing Chu, Huangqimei Zheng and Qianqian Wang
ISPRS Int. J. Geo-Inf. 2024, 13(7), 222; https://doi.org/10.3390/ijgi13070222 - 26 Jun 2024
Viewed by 1127
Abstract
Pansharpening is the fusion of panchromatic images and multispectral images to obtain images with high spatial resolution and high spectral resolution, which have a wide range of applications. At present, methods based on deep learning can fit the nonlinear features of images and [...] Read more.
Pansharpening is the fusion of panchromatic images and multispectral images to obtain images with high spatial resolution and high spectral resolution, which have a wide range of applications. At present, methods based on deep learning can fit the nonlinear features of images and achieve excellent image quality; however, the images generated with supervised learning approaches lack real-world applicability. Therefore, in this study, we propose an unsupervised pansharpening method based on a generative adversarial network. Considering the fine tubular structures in remote sensing images, a dense connection attention module is designed based on dynamic snake convolution to recover the details of spatial information. In the stage of image fusion, the fusion of features in groups is applied through the cross-scale attention fusion module. Moreover, skip layers are implemented at different scales to integrate significant information, thus improving the objective index values and visual appearance. The loss function contains four constraints, allowing the model to be effectively trained without reference images. The experimental results demonstrate that the proposed method outperforms other widely accepted state-of-the-art methods on the QuickBird and WorldView2 data sets. Full article
(This article belongs to the Special Issue Advances in AI-Driven Geospatial Analysis and Data Generation)
Show Figures

Figure 1

Figure 1
<p>Network architecture of UPGAN.</p>
Full article ">Figure 2
<p>Slender tubular structures in remote sensing images.</p>
Full article ">Figure 3
<p>The structure of DCAM.</p>
Full article ">Figure 4
<p>The framework of MSDFL.</p>
Full article ">Figure 5
<p>The structure of SSA.</p>
Full article ">Figure 6
<p>The channel attention mechanism in SSA.</p>
Full article ">Figure 7
<p>The framework of CSAF.</p>
Full article ">Figure 8
<p>Frameworks of the two discriminators: (<b>a</b>) spectral discriminator and (<b>b</b>) spatial discriminator.</p>
Full article ">Figure 9
<p>Qualitative comparison of UPGAN with 10 counterparts on a sample from the QB data set. (<b>a</b>) Brovey [<a href="#B14-ijgi-13-00222" class="html-bibr">14</a>]. (<b>b</b>) MTF_GLP_HPM [<a href="#B22-ijgi-13-00222" class="html-bibr">22</a>]. (<b>c</b>) PCA [<a href="#B13-ijgi-13-00222" class="html-bibr">13</a>]. (<b>d</b>) SFIM [<a href="#B15-ijgi-13-00222" class="html-bibr">15</a>]. (<b>e</b>) SR-D [<a href="#B29-ijgi-13-00222" class="html-bibr">29</a>]. (<b>f</b>) ZeRGAN [<a href="#B52-ijgi-13-00222" class="html-bibr">52</a>]. (<b>g</b>) UCGAN [<a href="#B51-ijgi-13-00222" class="html-bibr">51</a>]. (<b>h</b>) LDPNet [<a href="#B41-ijgi-13-00222" class="html-bibr">41</a>]. (<b>i</b>) PanGAN [<a href="#B47-ijgi-13-00222" class="html-bibr">47</a>]. (<b>j</b>) ZS-Pan [<a href="#B44-ijgi-13-00222" class="html-bibr">44</a>]. (<b>k</b>) UPGAN. (<b>l</b>) GT.</p>
Full article ">Figure 10
<p>The residual images between the pansharpened results and reference images in <a href="#ijgi-13-00222-f009" class="html-fig">Figure 9</a>. (<b>a</b>) Brovey [<a href="#B14-ijgi-13-00222" class="html-bibr">14</a>]. (<b>b</b>) MTF_GLP_HPM [<a href="#B22-ijgi-13-00222" class="html-bibr">22</a>]. (<b>c</b>) PCA [<a href="#B13-ijgi-13-00222" class="html-bibr">13</a>]. (<b>d</b>) SFIM [<a href="#B15-ijgi-13-00222" class="html-bibr">15</a>]. (<b>e</b>) SR-D [<a href="#B29-ijgi-13-00222" class="html-bibr">29</a>]. (<b>f</b>) ZeRGAN [<a href="#B52-ijgi-13-00222" class="html-bibr">52</a>]. (<b>g</b>) UCGAN [<a href="#B51-ijgi-13-00222" class="html-bibr">51</a>]. (<b>h</b>) LDPNet [<a href="#B41-ijgi-13-00222" class="html-bibr">41</a>]. (<b>i</b>) PanGAN [<a href="#B47-ijgi-13-00222" class="html-bibr">47</a>]. (<b>j</b>) ZS-Pan [<a href="#B44-ijgi-13-00222" class="html-bibr">44</a>]. (<b>k</b>) UPGAN. (<b>l</b>) GT.</p>
Full article ">Figure 11
<p>Qualitative comparison of UPGAN with 10 counterparts on a typical satellite image pair from the QB data set at full resolution. (<b>a</b>) Brovey [<a href="#B14-ijgi-13-00222" class="html-bibr">14</a>]. (<b>b</b>) MTF_GLP_HPM [<a href="#B22-ijgi-13-00222" class="html-bibr">22</a>]. (<b>c</b>) PCA [<a href="#B13-ijgi-13-00222" class="html-bibr">13</a>]. (<b>d</b>) SFIM [<a href="#B15-ijgi-13-00222" class="html-bibr">15</a>]. (<b>e</b>) SR-D [<a href="#B29-ijgi-13-00222" class="html-bibr">29</a>]. (<b>f</b>) ZeRGAN [<a href="#B52-ijgi-13-00222" class="html-bibr">52</a>]. (<b>g</b>) UCGAN [<a href="#B51-ijgi-13-00222" class="html-bibr">51</a>]. (<b>h</b>) LDPNet [<a href="#B41-ijgi-13-00222" class="html-bibr">41</a>]. (<b>i</b>) PanGAN [<a href="#B47-ijgi-13-00222" class="html-bibr">47</a>]. (<b>j</b>) ZS-Pan [<a href="#B44-ijgi-13-00222" class="html-bibr">44</a>]. (<b>k</b>) UPGAN. (<b>l</b>) PAN. (<b>m</b>) MS.</p>
Full article ">Figure 12
<p>Qualitative comparison of UPGAN with 10 counterparts on a sample from the WV2 data set. (<b>a</b>) Brovey [<a href="#B14-ijgi-13-00222" class="html-bibr">14</a>]. (<b>b</b>) MTF_GLP_HPM [<a href="#B22-ijgi-13-00222" class="html-bibr">22</a>]. (<b>c</b>) PCA [<a href="#B13-ijgi-13-00222" class="html-bibr">13</a>]. (<b>d</b>) SFIM [<a href="#B15-ijgi-13-00222" class="html-bibr">15</a>]. (<b>e</b>) SR-D [<a href="#B29-ijgi-13-00222" class="html-bibr">29</a>]. (<b>f</b>) ZeRGAN [<a href="#B52-ijgi-13-00222" class="html-bibr">52</a>]. (<b>g</b>) UCGAN [<a href="#B51-ijgi-13-00222" class="html-bibr">51</a>]. (<b>h</b>) LDPNet [<a href="#B41-ijgi-13-00222" class="html-bibr">41</a>]. (<b>i</b>) PanGAN [<a href="#B47-ijgi-13-00222" class="html-bibr">47</a>]. (<b>j</b>) ZS-Pan [<a href="#B44-ijgi-13-00222" class="html-bibr">44</a>]. (<b>k</b>) UPGAN. (<b>l</b>) GT.</p>
Full article ">Figure 13
<p>The residual images between the pansharpened results and reference images in <a href="#ijgi-13-00222-f012" class="html-fig">Figure 12</a>. (<b>a</b>) Brovey [<a href="#B14-ijgi-13-00222" class="html-bibr">14</a>]. (<b>b</b>) MTF_GLP_HPM [<a href="#B22-ijgi-13-00222" class="html-bibr">22</a>]. (<b>c</b>) PCA [<a href="#B13-ijgi-13-00222" class="html-bibr">13</a>]. (<b>d</b>) SFIM [<a href="#B15-ijgi-13-00222" class="html-bibr">15</a>]. (<b>e</b>) SR-D [<a href="#B29-ijgi-13-00222" class="html-bibr">29</a>]. (<b>f</b>) ZeRGAN [<a href="#B52-ijgi-13-00222" class="html-bibr">52</a>]. (<b>g</b>) UCGAN [<a href="#B51-ijgi-13-00222" class="html-bibr">51</a>]. (<b>h</b>) LDPNet [<a href="#B41-ijgi-13-00222" class="html-bibr">41</a>]. (<b>i</b>) PanGAN [<a href="#B47-ijgi-13-00222" class="html-bibr">47</a>]. (<b>j</b>) ZS-Pan [<a href="#B44-ijgi-13-00222" class="html-bibr">44</a>]. (<b>k</b>) UPGAN. (<b>l</b>) GT.</p>
Full article ">Figure 14
<p>Qualitative comparison of UPGAN with 10 counterparts on a typical satellite image pair from the QB data set at full resolution. (<b>a</b>) Brovey [<a href="#B14-ijgi-13-00222" class="html-bibr">14</a>]. (<b>b</b>) MTF_GLP_HPM [<a href="#B22-ijgi-13-00222" class="html-bibr">22</a>]. (<b>c</b>) PCA [<a href="#B13-ijgi-13-00222" class="html-bibr">13</a>]. (<b>d</b>) SFIM [<a href="#B15-ijgi-13-00222" class="html-bibr">15</a>]. (<b>e</b>) SR-D [<a href="#B29-ijgi-13-00222" class="html-bibr">29</a>]. (<b>f</b>) ZeRGAN [<a href="#B52-ijgi-13-00222" class="html-bibr">52</a>]. (<b>g</b>) UCGAN [<a href="#B51-ijgi-13-00222" class="html-bibr">51</a>]. (<b>h</b>) LDPNet [<a href="#B41-ijgi-13-00222" class="html-bibr">41</a>]. (<b>i</b>) PanGAN [<a href="#B47-ijgi-13-00222" class="html-bibr">47</a>]. (<b>j</b>) ZS-Pan [<a href="#B44-ijgi-13-00222" class="html-bibr">44</a>]. (<b>k</b>) UPGAN. (<b>l</b>) PAN. (<b>m</b>) MS.</p>
Full article ">Figure 15
<p>The connection structure in the generator.</p>
Full article ">Figure 16
<p>Average quantitative results for the number of sub-modules in DCAM on the QB data set. (<b>a</b>) PSNR. (<b>b</b>) SCC. (<b>c</b>) SAM. (<b>d</b>) ERGAS. (<b>e</b>) UIQI.</p>
Full article ">
20 pages, 21097 KiB  
Article
DPDU-Net: Double Prior Deep Unrolling Network for Pansharpening
by Yingxia Chen, Yuqi Li, Tingting Wang, Yan Chen and Faming Fang
Remote Sens. 2024, 16(12), 2141; https://doi.org/10.3390/rs16122141 - 13 Jun 2024
Viewed by 570
Abstract
The objective of the pansharpening task is to integrate multispectral (MS) images with low spatial resolution (LR) and to integrate panchromatic (PAN) images with high spatial resolution (HR) to generate HRMS images. Recently, deep learning-based pansharpening methods have been widely studied. However, traditional [...] Read more.
The objective of the pansharpening task is to integrate multispectral (MS) images with low spatial resolution (LR) and to integrate panchromatic (PAN) images with high spatial resolution (HR) to generate HRMS images. Recently, deep learning-based pansharpening methods have been widely studied. However, traditional deep learning methods lack transparency while deep unrolling methods have limited performance when using one implicit prior for HRMS images. To address this issue, we incorporate one implicit prior with a semi-implicit prior and propose a double prior deep unrolling network (DPDU-Net) for pansharpening. Specifically, we first formulate the objective function based on observation models of PAN and LRMS images and two priors of an HRMS image. In addition to the implicit prior in the image domain, we enforce the sparsity of the HRMS image in a certain multi-scale implicit space; thereby, the feature map can obtain better sparse representation ability. We optimize the proposed objective function via alternating iteration. Then, the iterative process is unrolled into an elaborate network, with each iteration corresponding to a stage of the network. We conduct both reduced-resolution and full-resolution experiments on two satellite datasets. Both visual comparisons and metric-based evaluations consistently demonstrate the superiority of the proposed DPDU-Net. Full article
Show Figures

Figure 1

Figure 1
<p>The overall framework of the proposed DPDU-Net. It iterates <span class="html-italic">T</span> stages, each consisting of a V-Net and a Z-Net.</p>
Full article ">Figure 2
<p>The detailed architecture of the transformation operator <math display="inline"><semantics> <msubsup> <mrow> <mo>{</mo> <msub> <mi>E</mi> <mi>l</mi> </msub> <mo>}</mo> </mrow> <mrow> <mi>l</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>L</mi> </msubsup> </semantics></math> and its inverse operation <math display="inline"><semantics> <msubsup> <mrow> <mo>{</mo> <msubsup> <mi>E</mi> <mi>l</mi> <mi>T</mi> </msubsup> <mo>}</mo> </mrow> <mrow> <mi>l</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>L</mi> </msubsup> </semantics></math>.</p>
Full article ">Figure 3
<p>The detailed architecture of the proposed ProxNet. For the <math display="inline"><semantics> <msub> <mi>t</mi> <mrow> <mi>t</mi> <mi>h</mi> </mrow> </msub> </semantics></math> iteration, the ProxNet takes <math display="inline"><semantics> <msup> <mrow> <mi mathvariant="bold-italic">Z</mi> </mrow> <mrow> <mi>t</mi> <mo>−</mo> <mfrac> <mn>1</mn> <mn>2</mn> </mfrac> </mrow> </msup> </semantics></math> as input and produces <math display="inline"><semantics> <msup> <mrow> <mi mathvariant="bold-italic">Z</mi> </mrow> <mi>t</mi> </msup> </semantics></math> as output.</p>
Full article ">Figure 4
<p>Visual results of qualitative comparisons on the reduced-resolution GF2 dataset. The regions of interest in the image, which were framed, have been enlarged and displayed in the lower left and lower right corners, respectively.</p>
Full article ">Figure 5
<p>Visual results of qualitative comparisons on the reduced-resolution WV3 dataset. The regions of interest in the image, which were framed, have been enlarged and displayed in the lower left and lower right corners, respectively.</p>
Full article ">Figure 6
<p>Error maps of qualitative comparisons on the reduced-resolution GF2 dataset.</p>
Full article ">Figure 7
<p>Error maps of qualitative comparisons on the reduced-resolution WV3 dataset.</p>
Full article ">Figure 8
<p>Visual results of qualitative comparisons on the full-resolution GF2 dataset. The regions of interest in the image, which were framed, have been enlarged and displayed in the lower left and lower right corners, respectively.</p>
Full article ">Figure 9
<p>Visual results of qualitative comparisons on the full-resolution WV3 dataset. The regions of interest in the image, which were framed, have been enlarged and displayed in the lower left and lower right corners, respectively.</p>
Full article ">Figure 10
<p>The line chart of the PSNR/SCC against scale. The vertical axis represents the evaluation indicators PSNR and SCC, with higher values indicating better performance, and the horizontal axis indicates the number of scales used in the CSC methods (0-scales indicates that the CSC method was performed directly in the image domain).</p>
Full article ">Figure 11
<p>The line chart of the PSNR/SCC against stage. The vertical axis represents the evaluation indicators PSNR and SCC, with higher values indicating better performance, and the horizontal axis indicates the number of stages in the deep unrolling network.</p>
Full article ">
21 pages, 60743 KiB  
Article
Deep Learning Hyperspectral Pansharpening on Large-Scale PRISMA Dataset
by Simone Zini, Mirko Paolo Barbato, Flavio Piccoli and Paolo Napoletano
Remote Sens. 2024, 16(12), 2079; https://doi.org/10.3390/rs16122079 - 8 Jun 2024
Cited by 1 | Viewed by 1526
Abstract
Hyperspectral pansharpening is crucial for the improvement of the usability of images in various applications. However, it remains underexplored due to a scarcity of data. The primary goal of pansharpening is to enhance the spatial resolution of hyperspectral images by reconstructing missing spectral [...] Read more.
Hyperspectral pansharpening is crucial for the improvement of the usability of images in various applications. However, it remains underexplored due to a scarcity of data. The primary goal of pansharpening is to enhance the spatial resolution of hyperspectral images by reconstructing missing spectral information without compromising consistency with the original data. This paper addresses the data gap by presenting a new hyperspectral dataset specifically designed for pansharpening and the evaluation of several deep learning strategies using this dataset. The new dataset has two crucial features that make it invaluable for deep learning hyperspectral pansharpening research. (1) It presents the highest cardinality of images in the state of the art, making it the first statistically relevant dataset for hyperspectral pansharpening evaluation, and (2) it includes a wide variety of scenes, ensuring robust generalization capabilities for various approaches. The data, collected by the ASI PRISMA satellite, cover about 262,200 km2 and their heterogeneity is ensured by a random sampling of the Earth’s surface. The analysis of the deep learning methods consists in the adaptation of these approaches to the PRISMA hyperspectral data and the quantitative and qualitative evaluation of their performance in this new scenario. The investigation included two settings: Reduced Resolution (RR) to evaluate the techniques in a controlled environment and Full Resolution (FR) for a real-world evaluation. In addition, for the sake of completeness, we have also included machine-learning-free approaches in both scenarios. Our comprehensive analysis reveals that data-driven neural network methods significantly outperform traditional approaches, demonstrating a superior adaptability and performance in hyperspectral pansharpening under both RR and FR protocols. Full article
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>Map of the patches acquired using the PRISMA satellite. On average, each patch covers about 1380 km<sup>2</sup> of soil. Each red box corresponds to a land-patch.</p>
Full article ">Figure 2
<p>Examples of PRISMA dataset entries, visualized in True Color RGB (641 nm, 563 nm, 478 nm). The <math display="inline"><semantics> <mrow> <mi>P</mi> <mi>A</mi> <mi>N</mi> </mrow> </semantics></math> image is at a resolution of 5 m per pixel, the <math display="inline"><semantics> <mrow> <mi>H</mi> <mi>S</mi> </mrow> </semantics></math> images are at 30 m per pixel, and the <math display="inline"><semantics> <mrow> <mi>H</mi> <msub> <mi>S</mi> <mo>↓</mo> </msub> </mrow> </semantics></math> are at 180 m per pixel.</p>
Full article ">Figure 3
<p>Distribution of the invalid bands for each PRISMA image collected. In (<b>a</b>), the scale indicates the percentage of invalid pixels for each band of each image collected from the PRISMA satellite. Bands that are considered invalid for at least one image (with more than 5% of the entries invalid) were selected for removal. In (<b>b</b>), the average spectral signatures per image and the bands excluded in the final version of the dataset are shown.</p>
Full article ">Figure 4
<p>Graph comparison of the results of the analyzed methods with the RR protocol. The larger the size of the circle, the higher the number of parameters (measured in millions).</p>
Full article ">Figure 5
<p>Pansharpening results on a <math display="inline"><semantics> <mrow> <mn>512</mn> <mo>×</mo> <mn>512</mn> </mrow> </semantics></math> tile of a test set image. For visualization purposes, the images have been linearly stretched between the 1st and 99th percentile of the image histogram. In the first row, images are visualized in True Color (641 nm, 563 nm, and 478 nm), and in the second row, the images are in False Color Composite (1586 nm, 1229 nm, and 770 nm).</p>
Full article ">Figure 6
<p>Pansharpening results on a <math display="inline"><semantics> <mrow> <mn>512</mn> <mo>×</mo> <mn>512</mn> </mrow> </semantics></math> tile of a test set image. For visualization purposes, the images have been linearly stretched between the 1st and 99th percentile of the image histogram. In the first row, images are visualized in True Color (641 nm, 563 nm, and 478 nm), and in the second row, the images are in False Color Composite (1586 nm, 1229 nm, and 770 nm).</p>
Full article ">Figure 7
<p>Pansharpening results on a <math display="inline"><semantics> <mrow> <mn>512</mn> <mo>×</mo> <mn>512</mn> </mrow> </semantics></math> tile of a test set image. For visualization purposes, the images have been linearly stretched between the 1st and 99th percentile of the image histogram. In the first row, the images are visualized in True Color (641 nm, 563 nm, and 478 nm), and in the second row, the images are in False Color Composite (1586 nm, 1229 nm, and 770 nm).</p>
Full article ">Figure 8
<p>Zoom of areas from the test images. Crops of dimension <math display="inline"><semantics> <mrow> <mn>128</mn> <mo>×</mo> <mn>128</mn> </mrow> </semantics></math>, at a resolution of 5 m/px, in True Color (641 nm, 563 nm, and 478 nm). As can be seen, repeated artifacts along the edges can be observed for the HySure method.</p>
Full article ">Figure 9
<p>Differences between the spectral signatures of the fused images with respect to the input image. The difference is evaluated as the average of the differences for each pixel of the five images reported in the row below the graph. The graph on the left shows the average spectral difference, while the graph on the right shows the difference normalized for each band.</p>
Full article ">Figure 9 Cont.
<p>Differences between the spectral signatures of the fused images with respect to the input image. The difference is evaluated as the average of the differences for each pixel of the five images reported in the row below the graph. The graph on the left shows the average spectral difference, while the graph on the right shows the difference normalized for each band.</p>
Full article ">Figure 10
<p>Spectral signatures obtained in six different areas, labeled as <span class="html-italic">Forest</span>, <span class="html-italic">Urban</span>, <span class="html-italic">Agriculture</span>, and <span class="html-italic">Water</span> areas. For each area, the spectral signatures of the input bands and those obtained by each pansharpening method are presented. The area used to extract the signatures is the one in the red box highlighted in each image. The image thumbnails are in True Color (641 nm, 563 nm, and 478 nm).</p>
Full article ">
22 pages, 7830 KiB  
Article
AIDB-Net: An Attention-Interactive Dual-Branch Convolutional Neural Network for Hyperspectral Pansharpening
by Qian Sun, Yu Sun and Chengsheng Pan
Remote Sens. 2024, 16(6), 1044; https://doi.org/10.3390/rs16061044 - 15 Mar 2024
Cited by 1 | Viewed by 1074
Abstract
Despite notable advancements achieved on Hyperspectral (HS) pansharpening tasks through deep learning techniques, previous methods are inherently constrained by convolution or self-attention intrinsic defects, leading to limited performance. In this paper, we proposed an Attention-Interactive Dual-Branch Convolutional Neural Network (AIDB-Net) for HS pansharpening. [...] Read more.
Despite notable advancements achieved on Hyperspectral (HS) pansharpening tasks through deep learning techniques, previous methods are inherently constrained by convolution or self-attention intrinsic defects, leading to limited performance. In this paper, we proposed an Attention-Interactive Dual-Branch Convolutional Neural Network (AIDB-Net) for HS pansharpening. Our model purely consists of convolutional layers and simultaneously inherits the strengths of both convolution and self-attention, especially the modeling of short- and long-range dependencies. Specially, we first extract, tokenize, and align the hyperspectral image (HSI) and panchromatic image (PAN) by Overlapping Patch Embedding Blocks. Then, we specialize a novel Spectral-Spatial Interactive Attention which is able to globally interact and fuse the cross-modality features. The resultant token-global similarity scores can guide the refinement and renewal of the textural details and spectral characteristics within HSI features. By deeply combined these two paradigms, our AIDB-Net significantly improve the pansharpening performance. Moreover, with the acceleration by the convolution inductive bias, our interactive attention can be trained without large scale dataset and achieves competitive time cost with its counterparts. Compared with the state-of-the-art methods, our AIDB-Net makes 5.2%, 3.1%, and 2.2% improvement on PSNR metric on three public datasets, respectively. Comprehensive experiments quantitatively and qualitatively demonstrate the effectiveness and superiority of our AIDB-Net. Full article
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>The overall architecture of our AIDB-Net adopts a dual-branch learning strategy. Specially, the HSI and PANs are first tokenized by the Overlapping Patch Embedding Blocks within the primary and subsidiary branches, respectively. Then, HSI and PAN features (tokens) are fused through the Spectral-Spatial Interactive Attention within the Residual Attention-Interactive Module, executed step by step. Finally, a convolutional layer, namely Residual Reconstruction Layer, is applied to recover and reconstruct the residual HSI. It’s note that the deep hyperspectral prior is utilized to provide suitable size HSI as input.</p>
Full article ">Figure 2
<p>The detailed structure of the Token-Mixer (<b>b</b>) and Spectral-Mixer (<b>c</b>). We repeat the operations in Token-Mixer in twice for better generation. Additionally, the Spectral Information Bottleneck (<b>d</b>) is introduced to extract the significant spectral characteristics.</p>
Full article ">Figure 3
<p>Visual results on the Pavia Centre dataset. (<b>a</b>) Reference. (<b>b</b>) PCA. (<b>c</b>) SFIM. (<b>d</b>) GS. (<b>e</b>) GSA. (<b>f</b>) CNMF. (<b>g</b>) MTF-GLP. (<b>h</b>) PanNet. (<b>i</b>) DBDENet. (<b>j</b>) DARN. (<b>k</b>) HyperKite. (<b>l</b>) AIDB-Net (Ours).</p>
Full article ">Figure 4
<p>The corresponding mean absolute error images on the Pavia Centre dataset. (<b>a</b>) Reference. (<b>b</b>) PCA. (<b>c</b>) SFIM. (<b>d</b>) GS. (<b>e</b>) GSA. (<b>f</b>) CNMF. (<b>g</b>) MTF-GLP. (<b>h</b>) PanNet. (<b>i</b>) DBDENet. (<b>j</b>) DARN. (<b>k</b>) HyperKite. (<b>l</b>) AIDB-Net (Ours).</p>
Full article ">Figure 5
<p>Comparison of spectral reflectance difference values of four randomly selected locations in <a href="#remotesensing-16-01044-f003" class="html-fig">Figure 3</a>. (<b>a</b>) Pixel (10,23). (<b>b</b>) Pixel (29,137). (<b>c</b>) Pixel(123,65). (<b>d</b>) Pixel (136,113).</p>
Full article ">Figure 6
<p>Visual results on the Botswana dataset. (<b>a</b>) Reference. (<b>b</b>) PCA. (<b>c</b>) SFIM. (<b>d</b>) GS. (<b>e</b>) GSA. (<b>f</b>) CNMF. (<b>g</b>) MTF-GLP. (<b>h</b>) PanNet. (<b>i</b>) DBDENet. (<b>j</b>) DARN. (<b>k</b>) HyperKite. (<b>l</b>) AIDB-Net (Ours).</p>
Full article ">Figure 7
<p>The corresponding mean absolute error images on the Botswana dataset. (<b>a</b>) Reference. (<b>b</b>) PCA. (<b>c</b>) SFIM. (<b>d</b>) GS. (<b>e</b>) GSA. (<b>f</b>) CNMF. (<b>g</b>) MTF-GLP. (<b>h</b>) PanNet. (<b>i</b>) DBDENet. (<b>j</b>) DARN. (<b>k</b>) HyperKite. (<b>l</b>) AIDB-Net (Ours).</p>
Full article ">Figure 8
<p>PSNR as a function of spectral band. (<b>a</b>) Pavia Centre dataset. (<b>b</b>) Botswana dataset. (<b>c</b>) Chikusei dataset.</p>
Full article ">Figure 9
<p>Visual results on the Chikusei dataset. (<b>a</b>) Reference. (<b>b</b>) PCA. (<b>c</b>) SFIM. (<b>d</b>) GS. (<b>e</b>) GSA. (<b>f</b>) CNMF. (<b>g</b>) MTF-GLP. (<b>h</b>) PanNet. (<b>i</b>) DBDENet. (<b>j</b>) DARN. (<b>k</b>) HyperKite. (<b>l</b>) AIDB-Net (Ours).</p>
Full article ">Figure 10
<p>The corresponding mean absolute error images on the Chikusei dataset. (<b>a</b>) Reference. (<b>b</b>) PCA. (<b>c</b>) SFIM. (<b>d</b>) GS. (<b>e</b>) GSA. (<b>f</b>) CNMF. (<b>g</b>) MTF-GLP. (<b>h</b>) PanNet. (<b>i</b>) DBDENet. (<b>j</b>) DARN. (<b>k</b>) HyperKite. (<b>l</b>) AIDB-Net (Ours).</p>
Full article ">Figure 11
<p>Three different learning strategies. (<b>a</b>) HSI-PAN. (<b>b</b>) PAN-HSI. (<b>c</b>) HSI+PAN.</p>
Full article ">Figure 12
<p>Visualization of the multi-group similarity interactive attention, when <math display="inline"><semantics> <mrow> <mi>G</mi> <mo>=</mo> <mn>8</mn> </mrow> </semantics></math>.</p>
Full article ">
28 pages, 11352 KiB  
Article
Pansharpening Low-Altitude Multispectral Images of Potato Plants Using a Generative Adversarial Network
by Sourav Modak, Jonathan Heil and Anthony Stein
Remote Sens. 2024, 16(5), 874; https://doi.org/10.3390/rs16050874 - 1 Mar 2024
Cited by 2 | Viewed by 2479
Abstract
Image preprocessing and fusion are commonly used for enhancing remote-sensing images, but the resulting images often lack useful spatial features. As the majority of research on image fusion has concentrated on the satellite domain, the image-fusion task for Unmanned Aerial Vehicle (UAV) images [...] Read more.
Image preprocessing and fusion are commonly used for enhancing remote-sensing images, but the resulting images often lack useful spatial features. As the majority of research on image fusion has concentrated on the satellite domain, the image-fusion task for Unmanned Aerial Vehicle (UAV) images has received minimal attention. This study investigated an image-improvement strategy by integrating image preprocessing and fusion tasks for UAV images. The goal is to improve spatial details and avoid color distortion in fused images. Techniques such as image denoising, sharpening, and Contrast Limited Adaptive Histogram Equalization (CLAHE) were used in the preprocessing step. The unsharp mask algorithm was used for image sharpening. Wiener and total variation denoising methods were used for image denoising. The image-fusion process was conducted in two steps: (1) fusing the spectral bands into one multispectral image and (2) pansharpening the panchromatic and multispectral images using the PanColorGAN model. The effectiveness of the proposed approach was evaluated using quantitative and qualitative assessment techniques, including no-reference image quality assessment (NR-IQA) metrics. In this experiment, the unsharp mask algorithm noticeably improved the spatial details of the pansharpened images. No preprocessing algorithm dramatically improved the color quality of the enhanced images. The proposed fusion approach improved the images without importing unnecessary blurring and color distortion issues. Full article
(This article belongs to the Special Issue Image Processing from Aerial and Satellite Imagery)
Show Figures

Graphical abstract

Graphical abstract
Full article ">Figure 1
<p>Sample images from the dataset: RGB, and spectral bands (NIR, red, red-edge, green). (<b>a</b>) RGB; (<b>b</b>) NIR; (<b>c</b>) red; (<b>d</b>) red-edge; (<b>e</b>) green.</p>
Full article ">Figure 2
<p>Integration of green channel, Near-Infrared (NIR), red, and red-edge channels into a composite multispectral image—combining multiple spectral layers for comprehensive data insight using the ArcPy Python module.</p>
Full article ">Figure 3
<p>PanColorGAN training: the pansharpened image <math display="inline"><semantics> <mover accent="true"> <msub> <mi>Y</mi> <mi>G</mi> </msub> <mo stretchy="false">^</mo> </mover> </semantics></math> is generated from the input <math display="inline"><semantics> <msub> <mi>X</mi> <mrow> <mi>G</mi> <mi>M</mi> <mi>S</mi> </mrow> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>X</mi> <mrow> <mi>M</mi> <mi>S</mi> </mrow> </msub> </semantics></math>. The quality of the generation was measured by calculating the reconstruction loss (<math display="inline"><semantics> <mrow> <mi>L</mi> <mi>o</mi> <mi>s</mi> <mi>s</mi> <mo>(</mo> <mi>L</mi> <mn>1</mn> <mo>)</mo> </mrow> </semantics></math>) between the colorized output <math display="inline"><semantics> <mover accent="true"> <msub> <mi>Y</mi> <mi>G</mi> </msub> <mo stretchy="false">^</mo> </mover> </semantics></math> and the multispectral input <math display="inline"><semantics> <msub> <mi>Y</mi> <mrow> <mi>M</mi> <mi>S</mi> </mrow> </msub> </semantics></math>. This loss serves as a crucial metric for training the PanColorGAN model. The methodology is adapted from the approach proposed in PanColorGAN paper [<a href="#B24-remotesensing-16-00874" class="html-bibr">24</a>].</p>
Full article ">Figure 4
<p>PanColorGAN architecture for training: grayscale <math display="inline"><semantics> <msub> <mi>X</mi> <mrow> <mi>P</mi> <mi>A</mi> <mi>N</mi> </mrow> </msub> </semantics></math> or <math display="inline"><semantics> <msub> <mi>X</mi> <mrow> <mi>G</mi> <mi>M</mi> <mi>S</mi> </mrow> </msub> </semantics></math> and multispectral <math display="inline"><semantics> <msub> <mi>X</mi> <mrow> <mi>M</mi> <mi>S</mi> </mrow> </msub> </semantics></math> are input to the generator <span class="html-italic">G</span>, which produces pansharpened <math display="inline"><semantics> <mover accent="true"> <msub> <mi>Y</mi> <mi>G</mi> </msub> <mo stretchy="false">^</mo> </mover> </semantics></math> or <math display="inline"><semantics> <mover accent="true"> <msub> <mi>Y</mi> <mi>P</mi> </msub> <mo stretchy="false">^</mo> </mover> </semantics></math>. In the discriminator network, <math display="inline"><semantics> <msub> <mi>X</mi> <mrow> <mi>G</mi> <mi>M</mi> <mi>S</mi> </mrow> </msub> </semantics></math>, <math display="inline"><semantics> <msub> <mi>X</mi> <mrow> <mi>M</mi> <mi>S</mi> </mrow> </msub> </semantics></math>, and <math display="inline"><semantics> <msub> <mi>Y</mi> <mrow> <mi>M</mi> <mi>S</mi> </mrow> </msub> </semantics></math> are included in the genuine batch image, while <math display="inline"><semantics> <mover accent="true"> <msub> <mi>Y</mi> <mi>P</mi> </msub> <mo stretchy="false">^</mo> </mover> </semantics></math> or <math display="inline"><semantics> <mover accent="true"> <msub> <mi>Y</mi> <mi>G</mi> </msub> <mo stretchy="false">^</mo> </mover> </semantics></math> are placed back into the fake batch image. The architecture is adapted from Ozcelik et al. (2021) [<a href="#B24-remotesensing-16-00874" class="html-bibr">24</a>].</p>
Full article ">Figure 5
<p>Pansharpening process: <math display="inline"><semantics> <msub> <mi>Y</mi> <mrow> <mi>P</mi> <mi>A</mi> <mi>N</mi> </mrow> </msub> </semantics></math> (750 × 750 pixels) and <math display="inline"><semantics> <msub> <mi>Y</mi> <mrow> <mi>M</mi> <mi>S</mi> </mrow> </msub> </semantics></math> (416 × 416 pixels) are resized into (<math display="inline"><semantics> <msubsup> <mi>X</mi> <mrow> <mi>P</mi> <mi>A</mi> <mi>N</mi> </mrow> <mrow> <mi>D</mi> <mi>O</mi> <mi>W</mi> <mi>N</mi> </mrow> </msubsup> </semantics></math>), (<math display="inline"><semantics> <msubsup> <mi>X</mi> <mrow> <mi>M</mi> <mi>S</mi> </mrow> <mrow> <mi>U</mi> <mi>P</mi> </mrow> </msubsup> </semantics></math>) of 512 × 512 pixel size and put into the generator network. The final enhanced pansharpened image is <math display="inline"><semantics> <mover accent="true"> <msub> <mi>Y</mi> <mrow> <mi>p</mi> <mi>s</mi> </mrow> </msub> <mo stretchy="false">^</mo> </mover> </semantics></math> (512 × 512 pixels).</p>
Full article ">Figure 6
<p>Visual representation of various image types: RGB (1), PAN (2), and spectral bands (green, NIR, red, red-edge) (3–6); Multispectral (MS) images (UN MS, MS WF, MS USM, MS TV, MS CL, MS CL WF, MS CL USM, MS CL TV) (7–14); and Pansharpened (PS) images (UN PS, PS WF, PS USM, PS TV, PS CL, PS CL WF, PS CL USM, PS CL TV) (15–22).</p>
Full article ">Figure 7
<p>Visual representation of various image types: RGB (1), PAN (2), and spectral bands (green, NIR, red, red-edge) (3–6); Multispectral (MS) images (UN MS, MS WF, MS USM, MS TV, MS CL, MS CL WF, MS CL USM, MS CL TV) (7–14); and Pansharpened (PS) images (UN PS, PS WF, PS USM, PS TV, PS CL, PS CL WF, PS CL USM, PS CL TV) (15–22).</p>
Full article ">Figure 8
<p>Visual comparison of (<b>A</b>) RGB, (<b>B</b>) PanColorGAN pansharpened and (<b>C</b>) Brovey transform-based pansharpened images, where the red circle in the RGB image highlights white flowers which disappear in both pansharpened images.</p>
Full article ">
Back to TopTop