Vegetation Greening for Winter Oblique Photography Using Cycle-Consistence Adversarial Networks
<p>A comparison of vegetation in winter oblique photography and summer oblique photography. (<b>a</b>) is winter oblique photography, and (<b>b</b>) is summer oblique photography.</p> "> Figure 2
<p>Structure of Cycle-consistence Adversarial Networks (CycleGAN).</p> "> Figure 3
<p>An example of “checkerboard artifacts” in the generated photography: (<b>a</b>) input photography; (<b>b</b>) parts of artifacts in generated photography; (<b>c</b>,<b>d</b>) are enlarged parts of (<b>b</b>) where artifacts are obvious.</p> "> Figure 3 Cont.
<p>An example of “checkerboard artifacts” in the generated photography: (<b>a</b>) input photography; (<b>b</b>) parts of artifacts in generated photography; (<b>c</b>,<b>d</b>) are enlarged parts of (<b>b</b>) where artifacts are obvious.</p> "> Figure 4
<p>Structure of residual block.</p> "> Figure 5
<p>Comparison of the 3D model of bad and good visual quality. (<b>a</b>) is the 3D model generated by the original photography and (<b>b</b>) is generated by the transferred photography.</p> "> Figure 6
<p>Results from different artifact reduction solutions in <a href="#sec3-symmetry-10-00294" class="html-sec">Section 3</a>. (<b>a</b>) is the input photography; (<b>b</b>) is the result of Solution (1); (<b>c</b>) is the result of Solution (2).</p> "> Figure 7
<p>Results of different generators. From left to right are: input, results of the generator of [<a href="#B30-symmetry-10-00294" class="html-bibr">30</a>], U-net generator and our modified generator.</p> "> Figure 8
<p>Results of different generators on another dataset. From left to right are the results from: input, the generator of [<a href="#B30-symmetry-10-00294" class="html-bibr">30</a>], the U-net generator and our modified generator.</p> "> Figure 9
<p>Results of different generators tested on another photograph set. From left to right are the results from: input, the U-net generator and our modified generator.</p> "> Figure 10
<p>(<b>a</b>) is the input, and (<b>b</b>) is the output.</p> "> Figure 11
<p>3D model generated by transferred photography from <a href="#symmetry-10-00294-f010" class="html-fig">Figure 10</a>b.</p> "> Figure 12
<p>(<b>a</b>) Input; (<b>b</b>) output.</p> ">
Abstract
:1. Introduction
- (1)
- Vegetation greening in winter oblique photography is achieved. In comparison with common methods, the infrared band is no longer required.
- (2)
- Checkerboard artifacts are reduced after CycleGAN is modified. The transferred photography can be applied in production.
- (3)
- The model can be trained with unpaired images, which is practical.
2. Related Works
3. The Proposed Method
3.1. CycleGAN
3.2. Adversarial Loss
3.3. Cycle Consistency Loss
3.4. Total Loss
3.5. Elimination of Checkerboard Artifacts
- (1)
- Renounce the use of the deconvolution operation. The method instead is: first, use up-sampling methods to build the image in the desired size; then, use the convolution operation to process the image. The choices of up-sampling methods are the nearest neighbor method and the bilinear method. The author of [29] recommended the nearest neighbor method.
- (2)
- Adjust the kernel’s size in the model’s generator. Adjust the kernel’s size to enable it to be split by stride. In CycleGAN’s generator, some layer’s kernel size is 3 with a stride of 2. Following the instruction of [29], these kernels’ size is modified to 4, so that it can be divided by 2.
Process 1: CycleGAN training process. |
Preparation: Training images of winter X and training images of summer Y, mapping G with generated parameters and mapping F with yielded parameters , discriminator DX with yielded parameters and discriminator DY with yielded parameters . Input: and Do Step1: update , to minimize LGAN(G,DY,X,Y) and . Step2: update , to minimize LGAN(F,DX,Y,X) and . Until convergence |
4. Experimental Results
4.1. Dataset
4.2. Implementation Details
4.3. Results and Comparison
5. Conclusions and Discussion
Author Contributions
Acknowledgments
Conflicts of Interest
References
- Wei, W.; Wenwen, H.; Jiao, Z. Pictometry Oblique Photography Technique and its application in 3D City Modeling. Geomat. Spat. Inf. Technol. 2011, 34, 181–183. [Google Scholar]
- Lou, B.; Geng, Z.X.; Wei, X.F.; Wang, L.F.; Li, Y.-J. Texture mapping of 3D city model based on Pictometry oblique image. Eng. Surv. Mapp. 2013, 22, 70–74. [Google Scholar] [CrossRef]
- Wang, L.; Sousa, W.P.; Peng, G.; Biging, G.S. Comparison of IKONOS and QuickBird images for mapping mangrove species on the Caribbean coast of Panama. Remote Sens. Environ. 2004, 91, 432–440. [Google Scholar] [CrossRef] [Green Version]
- Zhu, J.Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2242–2251. [Google Scholar]
- Hertzmann, A.; Jacobs, C.E.; Oliver, N.; Curless, B.; Salesin, D.H. Image Analogies. In Proceedings of the 28th Annual Conference on Computer Graphics and Interactive Techniques (SIGGRAPH’01), Los Angeles, CA, USA, 12–17 August 2001; pp. 327–340. [Google Scholar]
- Zhang, R.; Isola, P.; Efros, A.A. Colorful Image Colorization. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 649–666. [Google Scholar]
- Wang, X.; Gupta, A. Generative Image Modeling Using Style and Structure Adversarial Networks. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 318–335. [Google Scholar]
- Isola, P.; Zhu, J.Y.; Zhou, T.; Efros, A.A. Image-to-Image Translation with Conditional Adversarial Networks. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Honolulu, HI, USA, 21–26 July 2017; pp. 5967–5976. [Google Scholar]
- Johnson, J.; Alahi, A.; Li, F.F. Perceptual Losses for Real-Time Style Transfer and Super-Resolution. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 694–711. [Google Scholar]
- Laffont, P.Y.; Ren, Z.; Tao, X.; Qian, C.; Hays, J. Transient attributes for high-level understanding and editing of outdoor scenes. ACM Trans. Graph. 2014, 33, 1–11. [Google Scholar] [CrossRef]
- Rosales, R.; Achan, K.; Frey, B. Unsupervised Image Translation. In Proceedings of the IEEE International Conference on Computer Vision, Beijing, China, 13–16 October 2003; Volume 471, pp. 472–478. [Google Scholar]
- Kingma, D.P.; Welling, M. Auto-Encoding Variational Bayes. arXiv, 2013; arXiv:1312.6114. [Google Scholar]
- Liu, M.Y.; Breuel, T.; Kautz, J. Unsupervised Image-to-Image Translation Networks. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 8 December 2017. [Google Scholar]
- Liu, M.Y.; Tuzel, O. Coupled Generative Adversarial Networks. In Proceedings of the 30th Conference on Neural Information Processing Systems (NIPS 2016), Barcelona, Spain, 5–10 December 2016. [Google Scholar]
- Goodfellow, I.J.; Pougetabadie, J.; Mirza, M.; Xu, B.; Wardefarley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Networks. Adv. Neural Inf. Process. Syst. 2014, 3, 2672–2680. [Google Scholar]
- Denton, E.; Chintala, S.; Szlam, A.; Fergus, R. Deep Generative Image Models using a Laplacian Pyramid of Adversarial Networks. In Proceedings of the 29th Conference on Neural Information Processing Systems (NIPS 2015), Montreal, QC, Canada, 7–12 December 2015; pp. 1486–1494. [Google Scholar]
- Radford, A.; Metz, L.; Chintala, S. Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks. arXiv, 2015; arXiv:1511.06434. [Google Scholar]
- Im, D.J.; Kim, C.D.; Jiang, H.; Memisevic, R. Generating images with recurrent adversarial networks. arXiv, 2016; arXiv:1602.05110. [Google Scholar]
- Chen, X.; Duan, Y.; Houthooft, R.; Schulman, J.; Sutskever, I.; Abbeel, P. InfoGAN: Interpretable Representation Learning by Information Maximizing Generative Adversarial Nets. In Proceedings of the 30th Conference on Neural Information Processing Systems (NIPS 2016), Barcelona, Spain, 5–10 December 2016. [Google Scholar]
- Arjovsky, M.; Bottou, L. Towards Principled Methods for Training Generative Adversarial Networks. arXiv, 2017; arXiv:1701.04862. [Google Scholar]
- Zhao, J.; Mathieu, M.; Lecun, Y. Energy-based Generative Adversarial Network. arXiv, 2016; arXiv:1609.03126. [Google Scholar]
- Sundaram, N.; Brox, T.; Keutzer, K. Dense Point Trajectories by GPU-Accelerated Large Displacement Optical Flow; Springer: London, UK, 2010; pp. 438–451. [Google Scholar]
- Brislin, R.W. Back-translation for cross-cultural research. J. Cross Cult. Psychol. 1970, 1, 185–216. [Google Scholar] [CrossRef]
- Huang, Q.X.; Guibas, L. Consistent Shape Maps via Semidefinite Programming. Comput. Graph. Forum 2013, 32, 177–186. [Google Scholar] [CrossRef] [Green Version]
- Godard, C.; Aodha, O.M.; Brostow, G.J. Unsupervised Monocular Depth Estimation with Left-Right Consistency. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
- Zhou, T.; Krähenbühl, P.; Aubry, M.; Huang, Q.; Efros, A.A. Learning Dense Correspondence via 3D-Guided Cycle Consistency. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 117–126. [Google Scholar]
- Shi, W.; Caballero, J.; Theis, L.; Huszar, F.; Aitken, A.; Ledig, C.; Wang, Z. Is the deconvolution layer the same as a convolutional layer? arXiv, 2016; arXiv:1609.07009. [Google Scholar]
- Gauthier, J. Conditional Generative Adversarial Nets For Convolutional Face Generation. Class Project for Stanford CS231N: Convolutional Neural Networks for Visual Recognition, Winter Semester. 2014. Available online: http://www.foldl.me/uploads/2015/conditional-gans-face-generation/paper.pdf (accessed on 20 July 2018).
- Odena, A.; Dumoulin, V.; Olah, C. Deconvolution and Checkerboard Artifacts. Distill 2016, 1. [Google Scholar] [CrossRef] [Green Version]
- Dosovitskiy, A.; Brox, T. Generating Images with Perceptual Similarity Metrics based on Deep Networks. In Proceedings of the 30th Conference on Neural Information Processing Systems (NIPS 2016), Barcelona, Spain, 5–10 December 2016. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Ulyanov, D.; Vedaldi, A.; Lempitsky, V. Instance Normalization: The Missing Ingredient for Fast Stylization. arXiv, 2016; arXiv:1607.08022. [Google Scholar]
- Dan, C.C.; Giusti, A.; Gambardella, L.M.; Schmidhuber. Deep Neural Networks Segment Neuronal Membranes in Electron Microscopy Images. Adv. Neural Inf. Process. Syst. 2012, 25, 2852–2860. [Google Scholar]
- Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
Layer |
---|
Input |
Reflection padding (3 × 3) |
64 × 7 × 7 conv, Stride 1, Instance Norm, ReLU |
128 × 4 × 4 conv, Stride 2, Instance Norm, ReLU |
256 × 4 × 4 conv, Stride 2, Instance Norm, ReLU |
Residual Block, 256 filter (9 blocks) |
128 × 4 × 4 deconv, Stride 2, Instance Norm, ReLU |
64 × 4 × 4 deconv, Stride 2, Instance Norm, ReLU |
3 × 7 × 7 conv, stride 1 |
Tanh |
Generator from [30] | U-Net Generator | Our Modified Generator |
---|---|---|
0.7346 | 0.9189 | 0.8978 |
0.6279 | 0.8379 | 0.8363 |
Generator | Forward Cycle Loss | Backward Cycle Loss |
---|---|---|
Generator from [30] | 0.05 | 0.49 |
U-net generator | 0.11 | 0.55 |
Our modified generator | 0.015 | 0.32 |
© 2018 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
Xue, X.; Wu, C.; Sun, Z.; Wu, Y.; Xiong, N.N. Vegetation Greening for Winter Oblique Photography Using Cycle-Consistence Adversarial Networks. Symmetry 2018, 10, 294. https://doi.org/10.3390/sym10070294
Xue X, Wu C, Sun Z, Wu Y, Xiong NN. Vegetation Greening for Winter Oblique Photography Using Cycle-Consistence Adversarial Networks. Symmetry. 2018; 10(7):294. https://doi.org/10.3390/sym10070294
Chicago/Turabian StyleXue, Xiaowei, Chunxue Wu, Ze Sun, Yan Wu, and Neal N. Xiong. 2018. "Vegetation Greening for Winter Oblique Photography Using Cycle-Consistence Adversarial Networks" Symmetry 10, no. 7: 294. https://doi.org/10.3390/sym10070294
APA StyleXue, X., Wu, C., Sun, Z., Wu, Y., & Xiong, N. N. (2018). Vegetation Greening for Winter Oblique Photography Using Cycle-Consistence Adversarial Networks. Symmetry, 10(7), 294. https://doi.org/10.3390/sym10070294