Detection of Collapsed Buildings in Post-Earthquake Remote Sensing Images Based on the Improved YOLOv3
"> Figure 1
<p>Location of the study area.</p> "> Figure 2
<p>Labeled samples of collapsed buildings. The green rectangles in (<b>a</b>,<b>b</b>) are collapsed buildings.</p> "> Figure 3
<p>Image enhancement methods: (<b>a</b>) original image, (<b>b</b>) 90-degree rotation, (<b>c</b>) 180-degree rotation, (<b>d</b>) 270-degree rotation, (<b>e</b>) horizontal flip, (<b>f</b>) up-and-down flip, (<b>g</b>) color transformation, and (<b>h</b>) image stretching.</p> "> Figure 4
<p>Method flow chart.</p> "> Figure 5
<p>YOLOv3 network structure, where the blue and red lines represent two-fold up-sampling.</p> "> Figure 6
<p>Schematic of the YOLOv3 prediction bounding box with 13 cells × 13 cells. (<b>a</b>) YOLOv3 detection process on 13 cells × 13 cells feature map; (<b>b</b>) YOLOv3 detection result.</p> "> Figure 7
<p>Structure of the Darknet53 convolutional network.</p> "> Figure 8
<p>Basic units of the ShuffleNet (DWConv: depth-wise convolution; GConv: group convolution): (<b>a</b>) basic unit of ShuffleNet v1, (<b>b</b>) basic unit for scaling down in ShuffleNet v1, (<b>c</b>) basic unit of ShuffleNet v2, and (<b>d</b>) basic unit for scaling down in ShuffleNet v2.</p> "> Figure 9
<p>ShuffleNet v2 network structure</p> "> Figure 10
<p>Sigmoid function and its derivative.</p> "> Figure 11
<p>Three different overlapping situations for two rectangular boxes.</p> "> Figure 12
<p>Loss curves on the verification set of three YOLOv3 models.</p> "> Figure 13
<p>Comparison of the precision recall curves for the three YOLOv3 models for YOLOv3 in blue, YOLOv3-ShuffleNet in red, and YOLOv3-S-GIoU in green.</p> "> Figure 14
<p>Comparison of detection results of between the original YOLOv3 and YOLOv3-S-GIoU. The first column represents the original YOLOv3 detection results; the second column represents the YOLOv3-S-GIoU detection results; the third column denotes the images before the earthquake. (<b>a</b>,<b>b</b>,<b>d</b>,<b>e</b>,<b>g</b>,<b>h</b>) are the detection results of the images without adding noise; (<b>j</b>,<b>k</b>,<b>m</b>,<b>n</b>) are the detection results of the images after adding noise.</p> "> Figure 14 Cont.
<p>Comparison of detection results of between the original YOLOv3 and YOLOv3-S-GIoU. The first column represents the original YOLOv3 detection results; the second column represents the YOLOv3-S-GIoU detection results; the third column denotes the images before the earthquake. (<b>a</b>,<b>b</b>,<b>d</b>,<b>e</b>,<b>g</b>,<b>h</b>) are the detection results of the images without adding noise; (<b>j</b>,<b>k</b>,<b>m</b>,<b>n</b>) are the detection results of the images after adding noise.</p> "> Figure 15
<p>Images with noise: (<b>a</b>) original image, (<b>b</b>) image with Gaussian noise, and (<b>c</b>) image with salt-pepper noise.</p> "> Figure 16
<p>Comparison of the precision recall curves for the two YOLOv3 models after the addition of noise to the test set.</p> "> Figure 17
<p>Example of YOLOv3-S-GIoU model detection results. (<b>a</b>) Test image for the Yushu earthquake; (<b>b</b>) test image for the Wenchuan earthquake</p> "> Figure 17 Cont.
<p>Example of YOLOv3-S-GIoU model detection results. (<b>a</b>) Test image for the Yushu earthquake; (<b>b</b>) test image for the Wenchuan earthquake</p> ">
Abstract
:1. Introduction
2. Materials and Methods
2.1. Dataset
2.1.1. Remote Sensing Data Acquisition
2.1.2. Dataset Production
2.1.3. Dataset Enhancement
2.2. Method Flow
2.3. Improved YOLOv3 Network Structure
- Conclusion 1. When the feature channels of the convolution layer for the input and output are equal, the MAC (memory access cost) is the smallest, whereas the model speed is the fastest.
- Conclusion 2. Excessive grouping convolution will increase the MAC and slow down the model’s running speed.
- Conclusion 3. Fewer branches in the model results in a more rapid model running speed.
- Conclusion 4. The time consumption of the element-wise operations is much higher than that of the floating-point operations. Therefore, it is necessary to reduce the element-wise operations as much as possible.
2.4. Improved YOLOv3 Loss Function
3. Experimental Settings
3.1. Evaluation Indicators
3.1.1. Precision Recall Curve
3.1.2. Average Precision
3.1.3. F1 Score
3.1.4. FPS
3.2. Implement Environment and Model Training
4. Results
4.1. Quantitative Evaluation
4.2. PRC Evaluation
5. Discussion
6. Conclusions
Author Contributions
Funding
Acknowledgments
Conflicts of Interest
References
- Dell’Acqua, F.; Gamba, P. Remote sensing and earthquake damage assessment: Experiences, limits, and perspectives. Proc. IEEE 2012, 100, 2876–2890. [Google Scholar] [CrossRef]
- Chen, W. Research of Remote Sensing Application Technology Based on Earthquake Disaster Assessment; China Earthquake Administration Lanzhou Institute of Seismology: Lanzhou, China, 2007. [Google Scholar]
- Cooner, A.; Shao, Y.; Campbell, J. Detection of urban damage using remote sensing and machine learning algorithms: Revisiting the 2010 Haiti earthquake. Remote Sens. 2016, 8, 868. [Google Scholar] [CrossRef] [Green Version]
- Uprety, P.; Yamazaki, F.; Dell’Acqua, F. Damage detection using high-resolution SAR imagery in the 2009 L’Aquila, Italy, Earthquake. Earthq. Spectra 2013, 29, 1521–1535. [Google Scholar] [CrossRef] [Green Version]
- Menderes, A.; Erener, A.; Sarp, G. Automatic detection of damaged buildings after earthquake hazard by using remote sensing and information technologies. Procedia Earth Planet. Sci. 2015, 15, 257–262. [Google Scholar] [CrossRef] [Green Version]
- Gong, L.; Li, Q.; Zhang, J. Earthquake building damage detection with object-oriented change detection. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium-IGARSS, Melbourne, Australia, 21–26 July 2013; pp. 3674–3677. [Google Scholar]
- Ye, X.; Wang, J.; Qin, Q. Damaged building detection based on GF-1 satellite remote sensing image: A case study for Nepal MS8.1 earthquake. Acta Seismol. Sin. 2016, 38, 477–485. [Google Scholar]
- Dong, L.; Shan, J. A comprehensive review of earthquake-induced building damage detection with remote sensing techniques. ISPRS J. Photogramm. Remote Sens. 2013, 84, 85–99. [Google Scholar] [CrossRef]
- Janalipour, M.; Mohammadzadeh, A. A fuzzy-ga based decision making system for detecting damaged buildings from high-spatial resolution optical images. Remote Sens. 2017, 9, 349. [Google Scholar] [CrossRef] [Green Version]
- Dai, W.; Jin, L.; Li, G. Real-time airplane detection algorithm in remote-sensing images based on improved YOLOv3. Opto Electron. Eng. 2018, 45, 84–92. [Google Scholar]
- Girshick, R.; Donahue, J.; Darrell, T. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
- Ren, S.; He, K.; Girshick, R. Faster R-CNN: Towards real-time object detection with region proposal networks. In Proceedings of the Annual Conference on Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; pp. 91–99. [Google Scholar]
- He, K.; Gkioxari, G.; Dollár, P. Mask R-CNN. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2961–2969. [Google Scholar]
- Redmon, J.; Divvala, S.; Girshick, R. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
- Liu, W.; Anguelov, D.; Erhan, D. SSD: Single shot multibox detector. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 21–37. [Google Scholar]
- Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
- Redmon, J.; Farhadi, A. YOLOv3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
- Han, X.; Zhong, Y.; Zhang, L. An efficient and robust integrated geospatial object detection framework for high spatial resolution remote sensing imagery. Remote Sens. 2017, 9, 666. [Google Scholar] [CrossRef] [Green Version]
- Cheng, G.; Zhou, P.; Han, J. Learning rotation-invariant convolutional neural networks for object detection in VHR optical remote sensing images. IEEE Trans. Geosci. Remote Sens. 2016, 54, 7405–7415. [Google Scholar] [CrossRef]
- Zheng, Z.; Liu, Y.; Pan, C. Application of improved YOLOv3 in aircraft recognition of remote sensing images. Electron. Opt. Control. 2019, 26, 32–36. [Google Scholar]
- Duarte, D.; Nex, F.; Kerle, N. Satellite image classification of building damages using airborne and satellite image samples in a deep learning approach. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2018, 4, 89–96. [Google Scholar] [CrossRef] [Green Version]
- Ji, M.; Liu, L.; Buchroithner, M. Identifying collapsed buildings using post-earthquake satellite imagery and convolutional neural networks: A case study of the 2010 Haiti earthquake. Remote Sens. 2018, 10, 1689. [Google Scholar] [CrossRef] [Green Version]
- Everingham, M.; Van Gool, L.; Williams, C.K.I.; Winn, J.; Zisserman, A. The PASCAL visual object classes (VOC) challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar] [CrossRef] [Green Version]
- Perez, L.; Wang, J. The effectiveness of data augmentation in image classification using deep learning. arXiv 2017, arXiv:1712.04621. [Google Scholar]
- Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. arXiv 2015, arXiv:1502.03167. [Google Scholar]
- He, K.; Zhang, X.; Ren, S. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Ma, N.; Zhang, X.; Zheng, H. ShuffleNet v2: Practical guidelines for efficient CNN architecture design. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 116–131. [Google Scholar]
- Chollet, F. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1800–1807. [Google Scholar]
- Zhang, X.; Zhou, X.; Lin, M. ShuffleNet: An extremely efficient convolutional neural network for mobile devices. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6848–6856. [Google Scholar]
- Rezatofighi, H.; Tsoi, N.; Gwak, J.Y. Generalized intersection over union: A metric and a loss for bounding box regression. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Los Angeles, CA, USA, 16–19 June 2019; pp. 658–666. [Google Scholar]
- Tian, Y.; Yang, G.; Wang, Z. Apple detection during different growth stages in orchards using the improved YOLO-V3 model. Comput. Electron. Agric. 2019, 157, 417–426. [Google Scholar] [CrossRef]
- Benjdira, B.; Khursheed, T.; Koubaa, A. Car detection using unmanned aerial vehicles: Comparison between Faster R-CNN and YOLOv3. In Proceedings of the International Conference on Unmanned Vehicle Systems-Oman (UVS), Sultan Qaboos Univ, Muscat, Oman, 5–7 February 2019; pp. 1–6. [Google Scholar]
- Zhao, Y.; Ren, H.; Cao, D. The research of building earthquake damage object-oriented change detection based on ensemble classifier with remote sensing image. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium-IGARSS, Valencia, Spain, 22–27 July 2018; pp. 4950–4953. [Google Scholar]
- Wen, X.; Bi, X.; Xiang, W. Object-oriented collapsed building extraction from multi-source remote sensing imagery based on SVM. North China Earthq. Sci. 2015, 33, 13–19. [Google Scholar]
Number of Sample Images | Number of Collapsed Buildings | |
---|---|---|
Training set | 1456 | 8751 |
Validation set | 364 | 2516 |
Testing set | 360 | 2234 |
Ground Truth | ||
---|---|---|
Collapsed Building | Others | |
Collapsed building | True Positive (TP) | False Positive (FP) |
Others | False Negative (FN) | True Negative (TN) |
P (%) | R (%) | F1 (%) | AP (%) | FPS (f/s) | Parameter Size (M) | |
---|---|---|---|---|---|---|
YOLOv3 | 88 | 78 | 82.7 | 85.84 | 23.95 | 241 |
YOLOv3-ShuffleNet | 87 | 81 | 83.89 | 85.98 | 29.16 | 146 |
YOLOv3-S-GIoU | 93 | 88 | 90.43 | 90.89 | 29.23 | 146 |
P (%) | R (%) | F1 (%) | AP (%) | |
---|---|---|---|---|
YOLOv3 | 63 | 41 | 49.67 | 44.3 |
YOLOv3-S-GIoU | 86 | 74 | 79.55 | 79.8 |
© 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
Ma, H.; Liu, Y.; Ren, Y.; Yu, J. Detection of Collapsed Buildings in Post-Earthquake Remote Sensing Images Based on the Improved YOLOv3. Remote Sens. 2020, 12, 44. https://doi.org/10.3390/rs12010044
Ma H, Liu Y, Ren Y, Yu J. Detection of Collapsed Buildings in Post-Earthquake Remote Sensing Images Based on the Improved YOLOv3. Remote Sensing. 2020; 12(1):44. https://doi.org/10.3390/rs12010044
Chicago/Turabian StyleMa, Haojie, Yalan Liu, Yuhuan Ren, and Jingxian Yu. 2020. "Detection of Collapsed Buildings in Post-Earthquake Remote Sensing Images Based on the Improved YOLOv3" Remote Sensing 12, no. 1: 44. https://doi.org/10.3390/rs12010044
APA StyleMa, H., Liu, Y., Ren, Y., & Yu, J. (2020). Detection of Collapsed Buildings in Post-Earthquake Remote Sensing Images Based on the Improved YOLOv3. Remote Sensing, 12(1), 44. https://doi.org/10.3390/rs12010044