MDESNet: Multitask Difference-Enhanced Siamese Network for Building Change Detection in High-Resolution Remote Sensing Images
<p>Overview of the proposed MDESNet. The T0 and T1 remote sensing images were inputted into the Siamese network (based on ResNeSt-50), and bitemporal multiscale feature maps were obtained (resolutions are 1/4, 1/8, 1/16, and 1/32 of the input, respectively). Subsequently, these were, respectively, inputted into FPN to fully integrate the context information. The two semantic segmentation branches were decoded using the MSFF module, whereas the change detection branch used the FDE module to obtain multiscale change features, which were then decoded using the MSFF module. Finally, they were restored to the original image resolution using a 4× bilinear upsampling.</p> "> Figure 2
<p>Overview of the FDE module. The proposed FDE module has difference and concatenation branches. The former applies the sigmoid activation function to obtain the feature difference attention map after making a difference on bitemporal feature maps, whereas the latter element-wise multiplies the attention map and the concatenated features to enhance the difference. The feature maps of the four scales are operated in the same way.</p> "> Figure 3
<p>A diagram of the MSFF module. Its basic unit is FusionBlock composed of residual block and the scSE module. The feature maps of adjacent scales were successively passed through FusionBlock to output high-resolution fused feature maps. The MSFF module contained six FusionBlocks, and finally applied a 1 × 1 convolutional layer and softmax function to obtain the classification results.</p> "> Figure 4
<p>The images of 512 × 512 pixels are obtained by cropping the BCDD dataset. Each column represents a group of samples, containing five images of prechange, postchange, and ground truth. In T0 and T1 labels, white and black represent buildings and background, respectively. In change labels, white and black represent changed and unchanged areas, respectively. Panels (<b>a</b>–<b>e</b>) show changes in buildings, whereas (<b>f</b>,<b>g</b>) show no change.</p> "> Figure 5
<p>Visualization of ablation study on the BCDD dataset: (<b>a</b>,<b>b</b>) are bitemporal images; (<b>c</b>) the ground truth; (<b>d</b>) baseline; (<b>e</b>) baseline + FPN; (<b>f</b>) baseline + scSE; (<b>g</b>) baseline + FPN + scSE; (<b>h</b>) baseline + Seg; (<b>i</b>) baseline + Seg + FPN; (<b>j</b>) baseline + Seg + scSE; (<b>k</b>) baseline + Seg + FPN + scSE. Changed pixels are indicated by white, whereas unchanged areas are shown in black. Red and blue represent false and missed detections, respectively.</p> "> Figure 6
<p>(<b>1</b>) and (<b>2</b>) are two different examples; the difference being that buildings were also present in the pretemporal image in (<b>2</b>)(<b>a</b>). In the two groups, (<b>a</b>) and (<b>b</b>) show pre- and post-temporal images, respectively; (<b>c</b>) and (<b>d</b>) show the change ground truth and predicted result, respectively, where white represents the buildings and black represents the background. (<b>e</b>–<b>h</b>) Feature difference attention maps at 4 scales (16 × 16, 32 × 32, 64 × 64, and 128 × 128 pixels), in which blue to red represents enhancement from weak to strong.</p> "> Figure 7
<p>Visualized comparison of the results of several change detection methods on the BCDD dataset: (<b>a</b>,<b>b</b>) are bitemporal images; (<b>c</b>) is the ground truth; (<b>d</b>) FC-EF; (<b>e</b>) FC-Siam-conc; (<b>f</b>) FC-Siam-diff; (<b>g</b>) ChangeNet; (<b>h</b>) DASNet; (<b>i</b>) SNUNet-CD; (<b>j</b>) DTCDSCN(CD); (<b>k, ours</b>) MDESNet. Changed pixels are represented by white, whereas unchanged areas are shown in black.</p> "> Figure 8
<p>Comparison of F1-scores with and without semantic segmentation branches under different baseline models.</p> "> Figure 9
<p>The influence of β value on the performance of MDESNet.</p> "> Figure 10
<p>Comparison of the number of parameters between MDESNet, multitask model DTCDSCN, and 2 single-task models (change detection models and UNet++).</p> ">
Abstract
:1. Introduction
- We propose a multitask difference-enhanced Siamese network based on a fully convolutional structure, which consists of a main task branch for change detection and two auxiliary branches for extracting bitemporal buildings. The introduction of semantic constraints enabled the model to learn the features of targets, facilitating the avoidance of pseudo-changes. The MSFF module was designed as a decoder in three branches, and the scSE algorithm was introduced to improve the ability of the model to recover spatial details.
- We propose an FDE module that combines concatenation and differences. This module enhanced the differences in bitemporal features at different scales, increasing the distance between pairs of real-changing pixels and thus enlarging the interclass disparity. This improved its ability to detect real changes and its robustness to pseudo-changes.
- We verify the performance of the proposed method on BCDD and achieve the best F1-score (0.9124) compared with other baseline methods.
2. Materials and Methods
2.1. Related Work
- Siamese Network
- ResNeSt
- FPN
- scSE
2.2. Network Architecture
2.3. Feature Difference Enhancement Module
2.4. Multiscale Feature Fusion Module
2.5. Loss Function
3. Experiments and Results
3.1. Dataset
3.2. Experimental Details
3.2.1. Evaluation Metrics
3.2.2. Parameter Settings
3.3. Ablation Study
3.4. Comparative Study of Similarity Measures
- (a)
- Concatenation [27]: We concatenated the bitemporal features, and then used three consecutive 3 × 3 convolutional layers, which reduced the channels, to extract the change information from the connected features according to the FC-Siam-conc.
- (b)
- Difference [27]: We subtracted the bitemporal features from the corresponding channel dimension and used the absolute value of the difference as the changed feature.
- (c)
- Normalized difference [53]: Based on this difference, we performed a further normalization operation.
- (d)
- Local similarity attention module [31]: In this module, we extracted the similarity attention (SA) value from the input feature maps using the cosine distance and then multiplied the SA element by element with the bitemporal feature maps. Finally, we concatenated the bitemporal feature maps and applied a 3 × 3 convolutional layer to adjust the number of channels as the changed features.
3.5. Comparative Study Using Other Methods
- (a)
- FC-EF [27]: This model concatenated two temporal images and formed an image with skip connections and a U-shaped structure, using six channels as input. We extracted the change feature from the fused image, and finally obtained the change result using the softmax function.
- (b)
- FC-Siam-conc [27]: This model was an extension of FC-EF that used a Siamese network with the same structure and shared weights as the encoder. We concatenated the extracted bitemporal features and then inputted them into the decoder with skip connections to obtain the change results.
- (c)
- FC-Siam-diff [27]: This model was very similar to FC-Siam-conc; the difference was that we subtracted and obtained the absolute value of the extracted bitemporal features and then input them into the decoder with skip connections to obtain the change results.
- (d)
- ChangeNet [30]: This was proposed to detect changes in street scenes. We sampled the change features of the three scales to the same scale and used the softmax function to obtain the change results after summation, which located and identified the changes between image pairs.
- (e)
- DASNet [25]: Its core was set to use the spatial attention and channel attention algorithms to obtain more abundant discriminative features. Unlike other methods, this method gave a distance map as output and used a threshold algorithm to obtain the final changed results.
- (f)
- SNUNet-CD [54]: This method used a dense connection Siamese network, similar to UNet++, which is known to mitigate the effects of deep location information loss in neural networks. We employed the ensemble channel attention module (ECAM) to extract the representative features of different semantic levels and obtain the change results.
- (g)
- DTCDSCN [33]: This model was also a multitask Siamese network with two semantic segmentation branches and a change detection branch, similar to the proposed MDESNet. Its decoder structure was similar to that of FC-Siam-diff, except that it added the scSE module to improve the feature representation capabilities.
4. Discussion
4.1. The Effect of Semantic Segmentation Branches
4.2. The Effect of the Value of β in Loss Function
4.3. Comparison of the Number of Parameters and Prediction Efficiency
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
Abbreviations
MDESNet | Multitask difference-enhanced Siamese network |
FDE | Feature difference enhancement |
BCDD | Building change detection dataset |
PBCD | Pixel-based change detection |
OBCD | Object-based change detection |
CVA | Change vector analysis |
PCA | Principal component analysis |
FC-EF | Fully convolutional early fusion |
FC-Siam-conc | Fully convolutional Siamese-concatenation |
FC-Siam-diff | Fully convolutional Siamese-difference |
LSTM | Long short-term memory |
DASNet | Dual attentive fully convolutional Siamese network |
CDD | Change detection dataset |
UCD | Urban change detection dataset |
MSFF | Multiscale feature fusion |
ResNeSt | Split-attention networks |
FDM | Feature difference map |
FDAM | Feature difference attention map |
scSE | Concurrent spatial and channel squeeze and channel excitation |
OA | Overall accuracy |
TN | True negative |
TP | True positive |
FN | False negative |
FP | False positive |
DDP | Distributed data parallel |
SyncBN | Synchronized cross-GPU batch normalization |
Adam | Adaptive moment estimation |
SNUNet-CD | Siamese nested-UNet network for change detection |
ECAM | Ensemble channel attention module |
DTCDSCN | Dual-task constrained deep Siamese convolutional network |
PSPNet | Pyramid scene parsing network |
References
- Chen, Y.; Lin, Z.; Zhao, X.; Wang, G.; Gu, Y. Deep Learning-Based Classification of Hyperspectral Data. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2094–2107. [Google Scholar] [CrossRef]
- Melgani, F.; Bruzzone, L. Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1778–1790. [Google Scholar] [CrossRef] [Green Version]
- Tian, Y.; Wu, B.; Zhang, L.; Li, Q.; Jia, K.; Wen, M. Opium poppy monitoring with remote sensing in North Myanmar. Int. J. Drug Policy 2011, 22, 278–284. [Google Scholar] [CrossRef]
- Tian, Y.; Yin, K.; Lu, D.; Hua, L.; Zhao, Q.; Wen, M. Examining Land Use and Land Cover Spatiotemporal Change and Driving Forces in Beijing from 1978 to 2010. Remote Sens. 2014, 6, 10593–10611. [Google Scholar] [CrossRef] [Green Version]
- Yin, K.; Lu, D.; Tian, Y.; Zhao, Q.; Yuan, C. Evaluation of Carbon and Oxygen Balances in Urban Ecosystems Using Land Use/Land Cover and Statistical Data. Sustainability 2014, 7, 195–221. [Google Scholar] [CrossRef] [Green Version]
- Xue, J.; Xu, H.; Yang, H.; Wang, B.; Wu, P.; Choi, J.; Cai, L.; Wu, Y. Multi-Feature Enhanced Building Change Detection Based on Semantic Information Guidance. Remote Sens. 2021, 13, 4171. [Google Scholar] [CrossRef]
- Wang, H.; Lv, X.; Zhang, K.; Guo, B. Building Change Detection Based on 3D Co-Segmentation Using Satellite Stereo Imagery. Remote Sens. 2022, 14, 628. [Google Scholar] [CrossRef]
- Bruzzone, L.; Bovolo, F. A Novel Framework for the Design of Change-Detection Systems for Very-High-Resolution Remote Sensing Images. Proc. IEEE 2013, 101, 609–630. [Google Scholar] [CrossRef]
- Dell’Acqua, F.; Gamba, P.; Ferrari, A.; Palmason, J.A.; Benediktsson, J.A.; Arnason, K. Exploiting spectral and spatial information in hyperspectral urban data with high resolution. IEEE Geosci. Remote Sens. Lett. 2004, 1, 322–326. [Google Scholar] [CrossRef]
- Gong, P.; Li, X.; Xu, B. Interpretation Theory and Application Method Development for Information Extraction from High Resolution Remotely Sensed Data. J. Remote Sens. 2006, 10, 1–5. [Google Scholar]
- Myint, S.W.; Lam, N.S.N.; Tyler, J.M. Wavelets for urban spatial feature discrimination: Comparisons with fractal, spatial autocorrelation, and spatial co-occurrence approaches. Photogramm. Eng. Remote Sens. 2004, 70, 803–812. [Google Scholar] [CrossRef] [Green Version]
- Weismiller, R.A.; Kristof, S.J.; Scholz, D.K.; Anuta, P.E.; Momin, S.A. Change detection in coastal zone environment. Photogramm. Eng. Remote Sens. 1978, 12, 1533–1539. [Google Scholar]
- Zhang, Z.; Li, Z.; Tian, X. Vegetation change detection research of Dunhuang city based on GF-1 data. Int. J. Coal Sci. Technol. 2018, 5, 105–111. [Google Scholar] [CrossRef] [Green Version]
- Singh, A. Review Article Digital change detection techniques using remotely-sensed data. Int. J. Remote Sens. 1989, 10, 989–1003. [Google Scholar] [CrossRef] [Green Version]
- Bruzzone, L.; Prieto, D.F. Automatic analysis of the difference image for unsupervised change detection. IEEE Trans. Geosci. Remote Sens. 2000, 38, 1171–1182. [Google Scholar] [CrossRef] [Green Version]
- Celik, T. Unsupervised Change Detection in Satellite Images Using Principal Component Analysis and K-Means Clustering. IEEE Geosci. Remote Sens. Lett. 2009, 6, 772–776. [Google Scholar] [CrossRef]
- Li, P.; Xu, H. Land-cover change detection using one-class support vector machine. Photogramm. Eng. Remote Sens. 2010, 76, 255–263. [Google Scholar] [CrossRef]
- Seo, D.; Kim, Y.; Eo, Y.; Park, W.; Park, H. Generation of Radiometric, Phenological Normalized Image Based on Random Forest Regression for Change Detection. Remote Sens. 2017, 9, 1163. [Google Scholar] [CrossRef] [Green Version]
- Lyu, H.; Lu, H.; Mou, L. Learning a Transferable Change Rule from a Recurrent Neural Network for Land Cover Change Detection. Remote Sens. 2016, 8, 506. [Google Scholar] [CrossRef] [Green Version]
- Hay, G.J.; Niemann, K.O. Visualizing 3-D texture: A three-dimensional structural approach to model forest texture. Can. J. Remote Sens. 1994, 20, 90–101. [Google Scholar]
- Lefebvre, A.; Corpetti, T.; Hubert-Moy, L. Object-Oriented Approach and Texture Analysis for Change Detection in Very High Resolution Images. In Proceedings of the 2008 IEEE International Geoscience and Remote Sensing Symposium, Boston, MA, USA, 6–11 July 2008; pp. 663–666. [Google Scholar]
- Yu, W.; Zhou, W.; Qian, Y.; Yan, J. A new approach for land cover classification and change analysis: Integrating backdating and an object-based method. Remote Sens. Environ. 2016, 177, 37–47. [Google Scholar] [CrossRef]
- Zhou, W.; Troy, A.; Grove, M.J.S. Object-based Land Cover Classification and Change Analysis in the Baltimore Metropolitan Area Using Multitemporal High Resolution Remote Sensing Data. Sensors 2008, 8, 1613–1636. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Jung, S.; Lee, W.H.; Han, Y. Change Detection of Building Objects in High-Resolution Single-Sensor and Multi-Sensor Imagery Considering the Sun and Sensor’s Elevation and Azimuth Angles. Remote Sens. 2021, 13, 3660. [Google Scholar] [CrossRef]
- Chen, J.; Yuan, Z.; Peng, J.; Chen, L.; Huang, H.; Zhu, J.; Liu, Y.; Li, H. DASNet: Dual Attentive Fully Convolutional Siamese Networks for Change Detection in High-Resolution Satellite Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 1194–1206. [Google Scholar] [CrossRef]
- LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
- Daudt, R.C.; Saux, B.L.; Boulch, A. Fully Convolutional Siamese Networks for Change Detection. arXiv 2018, arXiv:1810.08462. [Google Scholar]
- Papadomanolaki, M.; Verma, S.; Vakalopoulou, M.; Gupta, S.; Karantzalos, K. Detecting urban changes with recurrent neural networks from multitemporal Sentinel-2 data. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Yokohama, Japan, 28 July–1 August 2019. [Google Scholar]
- Li, S.; Huo, L. Remote Sensing Image Change Detection Based on Fully Convolutional Network With Pyramid Attention. In Proceedings of the 2021 IEEE International Geoscience and Remote Sensing Symposium IGARSS, Brussels, Belgium, 11–16 July 2021; pp. 4352–4355. [Google Scholar]
- Varghese, A.; Gubbi, J.; Ramaswamy, A.; Balamuralidhar, P. ChangeNet: A Deep Learning Architecture for Visual Change Detection. In Proceedings of the Computer Vision—ECCV 2018 Workshops; Springer: Cham, Switzerland, 2019; pp. 129–145. [Google Scholar]
- Lee, H.; Lee, K.; Kim, J.H.; Na, Y.; Park, J.; Choi, J.P.; Hwang, J.Y. Local Similarity Siamese Network for Urban Land Change Detection on Remote Sensing Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 4139–4149. [Google Scholar] [CrossRef]
- Li, X.; He, M.; Li, H.; Shen, H. A Combined Loss-Based Multiscale Fully Convolutional Network for High-Resolution Remote Sensing Image Change Detection. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
- Liu, Y.; Pang, C.; Zhan, Z.; Zhang, X.; Yang, X. Building Change Detection for Remote Sensing Images Using a Dual Task Constrained Deep Siamese Convolutional Network Model. arXiv 2019, arXiv:1909.07726. [Google Scholar] [CrossRef]
- Lin, T.; Dollar, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature Pyramid Networks for Object Detection. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 936–944. [Google Scholar]
- Roy, A.G.; Navab, N.; Wachinger, C. Concurrent Spatial and Channel ’Squeeze & Excitation’ in Fully Convolutional Networks. arXiv 2018, arXiv:1803.02579. [Google Scholar]
- Bromley, J.; Guyon, I.; LeCun, Y.; Säckinger, E.; Shah, R. Signature verification using a “Siamese” time delay neural network. In Proceedings of the Proceedings of the 6th International Conference on Neural Information Processing Systems, Denver, CO, USA, 29 November–2 December 1993; pp. 737–744. [Google Scholar]
- Chopra, S.; Hadsell, R.; LeCun, Y. Learning a similarity metric discriminatively, with application to face verification. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Diego, CA, USA, 20–25 June 2005; pp. 539–546. [Google Scholar]
- Benajiba, Y.; Sun, J.; Zhang, Y.; Jiang, L.; Weng, Z.; Biran, O. Siamese Networks for Semantic Pattern Similarity. arXiv 2018, arXiv:1812.06604. [Google Scholar]
- Ranasinghe, T.; Orasan, C.; Mitkov, R. Semantic Textual Similarity with Siamese Neural Networks. In Proceedings of the International Conference on Recent Advances in Natural Language Processing, Varna, Bulgaria, 2–4 September 2019; pp. 1004–1011. [Google Scholar]
- Zhang, H.; Wu, C.; Zhang, Z.; Zhu, Y.; Zhang, Z.-L.; Lin, H.; Sun, Y.; He, T.; Mueller, J.; Manmatha, R.; et al. ResNeSt: Split-Attention Networks. arXiv 2020, arXiv:2004.08955. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Xie, S.; Girshick, R.; Dollár, P.; Tu, Z.; He, K. Aggregated Residual Transformations for Deep Neural Networks. arXiv 2016, arXiv:1611.05431. [Google Scholar]
- Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. arXiv 2017, arXiv:1709.01507. [Google Scholar]
- Kirillov, A.; Girshick, R.; He, K. Panoptic Feature Pyramid Networks. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 6392–6401. [Google Scholar]
- Zheng, Z.; Zhong, Y.; Wang, J. Foreground-Aware Relation Network for Geospatial Object Segmentation in High Spatial Resolution Remote Sensing Imagery. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 4095–4104. [Google Scholar]
- Zhang, X.; He, L.; Qin, K.; Dang, Q.; Si, H.; Tang, X.; Jiao, L. SMD-Net: Siamese Multi-Scale Difference-Enhancement Network for Change Detection in Remote Sensing. Remote Sens. 2022, 14, 1580. [Google Scholar] [CrossRef]
- Zheng, D.; Wei, Z.; Wu, Z.; Liu, J. Learning Pairwise Potential CRFs in Deep Siamese Network for Change Detection. Remote Sens. 2022, 14, 841. [Google Scholar] [CrossRef]
- Lin, T.; Goyal, P.; Girshick, R.; He, K.; Dollar, P. Focal Loss for Dense Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 318–327. [Google Scholar] [CrossRef] [Green Version]
- Ji, S.; Wei, S.; Lu, M. Fully Convolutional Networks for Multisource Building Extraction From an Open Aerial and Satellite Imagery Data Set. IEEE Trans. Geosci. Remote Sens. 2019, 57, 574–586. [Google Scholar] [CrossRef]
- Zhang, H.; Dana, K.; Shi, J.; Zhang, Z.; Wang, X.; Tyagi, A.; Agrawal, A. Context Encoding for Semantic Segmentation. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7151–7160. [Google Scholar]
- Kingma, D.P.; Ba, J.J.A. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
- Loshchilov, I.; ArXiv, F.H.J. SGDR: Stochastic Gradient Descent with Warm Restarts. arXiv 2017, arXiv:1608.03983. [Google Scholar]
- Zhang, M.; Shi, W. A Feature Difference Convolutional Neural Network-Based Change Detection Method. IEEE Trans. Geosci. Remote Sens. 2020, 58, 7232–7246. [Google Scholar] [CrossRef]
- Fang, S.; Li, K.; Shao, J.; Li, Z. SNUNet-CD: A Densely Connected Siamese Network for Change Detection of VHR Images. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
- Olaf, R.; Philipp, F.; Thomas, B. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer Assisted Intervention, Munich, Germany, 5–9 October 2015. [Google Scholar]
- Zhou, Z.; Md Mahfuzur Rahman, S.; Nima, T.; Jianming, L. UNet++: A Nested U-Net Architecture for Medical Image Segmentation. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support, Proceedings of the DLMIA: International Workshop on Deep Learning in Medical Image Analysis, Granada, Spain, 20 September 2018; Springer: Berlin/Heidelberg, Germany, 2018. [Google Scholar]
- Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid Scene Parsing Network. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–16 July 2017; pp. 6230–6239. [Google Scholar]
- Liang-Chieh, C.; Yukun, Z.; George, P.; Florian, S.; Hartwig, A. Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
Clear Data | Changed Pixels | Unchanged Pixels | C/UC |
---|---|---|---|
No | 21,188,729 | 450,670,471 | 21.2693 |
Yes | 21,188,729 | 240,958,226 | 11.3720 |
Method | Seg | FPN | scSE | F1 (cd) | F1 (seg) |
---|---|---|---|---|---|
Baseline | 0.7791 | - | |||
Baseline + FPN | √ | 0.8857 | - | ||
Baseline + scSE | √ | 0.8395 | - | ||
Baseline + FPN + scSE | √ | √ | 0.8863 | - | |
Baseline + Seg | √ | 0.8149 | 0.9059 | ||
Baseline + Seg + FPN | √ | √ | 0.9032 | 0.9195 | |
Baseline + Seg + scSE | √ | √ | 0.8774 | 0.9297 | |
Baseline + Seg + FPN + scSE | √ | √ | √ | 0.9124 | 0.9441 |
Method | OA | Pre. | Rec. | F1 |
---|---|---|---|---|
Concatenate | 0.9814 | 0.8458 | 0.9100 | 0.8767 |
Difference | 0.9848 | 0.8750 | 0.9231 | 0.8984 |
Normalized difference | 0.9772 | 0.8810 | 0.9559 | 0.8974 |
Local similarity attention | 0.9807 | 0.8111 | 0.9571 | 0.8781 |
FDE (ours) | 0.9874 | 0.9264 | 0.8988 | 0.9124 |
Method | OA | Pre. | Rec. | F1 |
---|---|---|---|---|
FC-EF | 0.9747 | 0.8553 | 0.7943 | 0.8237 |
FC-Siam-conc | 0.9662 | 0.7199 | 0.8932 | 0.7972 |
FC-Siam-diff | 0.9555 | 0.6425 | 0.9056 | 0.7517 |
ChangeNet | 0.9378 | 0.5560 | 0.8119 | 0.6600 |
DASNet | 0.9802 | 0.8430 | 0.9266 | 0.8828 |
SNUNet-CD | 0.9792 | 0.8675 | 0.8494 | 0.8584 |
DTCDSCN | 0.9717 | 0.7469 | 0.9233 | 0.8258 |
MDESNet (ours) | 0.9874 | 0.9264 | 0.8988 | 0.9124 |
Method | OA | Pre. | Rec. | F1 |
---|---|---|---|---|
UNet | 0.9801 | 0.9546 | 0.9410 | 0.9478 |
UNet++ | 0.9814 | 0.9469 | 0.9568 | 0.9518 |
PSPNet | 0.9730 | 0.9258 | 0.9341 | 0.9299 |
DeepLabv3+ | 0.9808 | 0.9519 | 0.9477 | 0.9498 |
FarSeg | 0.9816 | 0.9587 | 0.9447 | 0.9516 |
DTCDSCN | 0.9711 | 0.9158 | 0.9352 | 0.9255 |
MDESNet (ours) | 0.9792 | 0.9485 | 0.9397 | 0.9441 |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Zheng, J.; Tian, Y.; Yuan, C.; Yin, K.; Zhang, F.; Chen, F.; Chen, Q. MDESNet: Multitask Difference-Enhanced Siamese Network for Building Change Detection in High-Resolution Remote Sensing Images. Remote Sens. 2022, 14, 3775. https://doi.org/10.3390/rs14153775
Zheng J, Tian Y, Yuan C, Yin K, Zhang F, Chen F, Chen Q. MDESNet: Multitask Difference-Enhanced Siamese Network for Building Change Detection in High-Resolution Remote Sensing Images. Remote Sensing. 2022; 14(15):3775. https://doi.org/10.3390/rs14153775
Chicago/Turabian StyleZheng, Jiaxiang, Yichen Tian, Chao Yuan, Kai Yin, Feifei Zhang, Fangmiao Chen, and Qiang Chen. 2022. "MDESNet: Multitask Difference-Enhanced Siamese Network for Building Change Detection in High-Resolution Remote Sensing Images" Remote Sensing 14, no. 15: 3775. https://doi.org/10.3390/rs14153775