Paper 3
Paper 3
Zhengdong Li
Department of Electronic and Computer Engineering, HKUST
Hong Kong S.A.R, CHINA
zlifd@connect.ust.hk
arXiv:2412.19479v1 [cs.CV] 27 Dec 2024
Abstract
Table 3. Evaluation results for PSNR, SSIM and mean time for
deblurring on single image in second on the trained GAN model.
- PSNR ↑ SSIM ↑
this work 29.16 0.75
[4] 28.7 0.958
[5] 29.55 0.934
[7] 28.93 0.91
future work may include but not limited to, (1) construct a
deeper neural network architecture, like more convolutional
layers and parameters such that the model can better learn
the features during training, (2) train with a larger dataset,
like [3] and [10] which comprises over 70 real world videos
Figure 4. Sample output using blurry image from GoPro dataset
captured in order to generate more training images, (3) work
in validation part.
on CUDA GPU environment to shorten the training time,
(4) explore different architectures such as U-Net [8], which
performance. For example, the edges of the whole building have shown promise in image-to-image translation tasks,
are sharper after deblurred by the GAN model. In Fig. 6, capturing more complex features and improving the qual-
they are the self-proposed motion-blur images in both in- ity of the generated images, and (5) leverage a pre-trained
door and outdoor, which also yield good deblur quality, and model as backbone which has already learned useful fea-
this can be clearly seen in the edges of the wall in the mid- tures from a large dataset can accelerate training and im-
dle part of Fig. 6. However, the deblur performance is not prove performance.
that significant in the self-proposed blurry images when it For our potential application, we aim to utilize a pre-
does not have enough motion-blur pixels from the input im- trained GAN model to deblur motion-blurred images cap-
ages, such as in Fig. 5. One reason could be the input image tured by smartphones. This innovative approach seeks
is not fully motion-blurred, which makes it challenging for to address the common issue of poor-quality images that
the Generator to produce a clear image based on the training arise from hand vibrations, particularly when capturing fast-
from the GoPro dataset. Hence, this is also another reason moving subjects. By leveraging the capabilities of GANs,
why we do not have a higher SSIM in the previous met- we hope to enhance image clarity and restore details that
rics evaluation part since not all the input images are truly are typically lost during motion blur. However, a significant
motion-blur. challenge we face is the difficulty in obtaining authentic
motion-blurred images from smartphones. Many modern
5. Future Work devices are equipped with advanced optical image stabiliza-
tion (OIS) [6] features that effectively mitigate motion blur,
We aim to improve the deblur performance for the GAN resulting in clearer images even in shaky conditions. This
model, yielding higher SSIM and PSNR for evaluation met- technological advancement, while beneficial for everyday
rics and produce sharper images per pixels for for graphical photography, poses a hurdle for our research, as it limits the
representations in the future. Hence, the direction of the availability of suitable training data for our GAN model.
blurred versions. This dataset provided an ideal founda-
tion for our adversarial training approach, which consists
of a Generator and a Discriminator competing with each
other. Through this dynamic interaction, the model progres-
sively enhances the quality of the generated images, ulti-
mately leading to more realistic and visually appealing out-
put. Throughout our experiments, we achieved a notable
mean PSNR of 29.1644 and a mean SSIM of 0.7459, indi-
cating a significant enhancement in image clarity and struc-
tural fidelity compared to the original blurred images. Ad-
ditionally, the average deblurring time of 4.6921 seconds
demonstrates the efficiency of our approach in real-world
applications. The results obtained from our GAN model
highlight the effectiveness of using adversarial training for
image restoration. The output images exhibit sharper details
and improved clarity, showcasing the potential of GANs in
tackling real-world challenges associated with motion blur.
This work not only underscores the advancements in image
processing techniques but also opens avenues for further
research and development in the field of computer vision,
where GANs can be applied to various image enhancement
tasks. Overall, our findings contribute to the growing body
of knowledge surrounding generative models and their prac-
tical applications in restoring image quality.
References
[1] Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing
Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and
Yoshua Bengio. Generative adversarial nets. Advances in
neural information processing systems, 27, 2014. 1
Figure 6. Another sample output with self-proposed blurry im- [2] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun.
ages. Since input images have more blurry pixels, so the output Deep residual learning for image recognition. In Proceed-
images have better deblurred quality. For instance, the edge of the ings of the IEEE conference on computer vision and pattern
wall is more clear than before in the middle of the figure. recognition, pages 770–778, 2016. 2
[3] Hamed Kiani Galoogahi, Ashton Fagg, Chen Huang, Deva
Ramanan, and Simon Lucey. Need for speed: A bench-
As shown in Fig. 5, it is somehow difficult to obtain some mark for higher frame rate object tracking. In Proceedings of
motion-blur images. Bad input quality in blurry images re- the IEEE international conference on computer vision, pages
sults in bad output quality in deblurred images as well. 1125–1134, 2017. 4
To overcome this obstacle, using existing datasets that [4] Orest Kupyn, Volodymyr Budzan, Mykola Mykhailych,
Dmytro Mishkin, and Jiřı́ Matas. Deblurgan: Blind mo-
contain examples of blurred images or simulating motion
tion deblurring using conditional adversarial networks. In
blur in controlled environments could be one of the solu-
Proceedings of the IEEE conference on computer vision and
tions. By addressing these challenges, we aim to develop pattern recognition, pages 8183–8192, 2018. 1, 2, 3, 4
a robust solution that enhances the quality of smartphone [5] Orest Kupyn, Tetiana Martyniuk, Junru Wu, and Zhangyang
photography, ultimately providing users with clearer, more Wang. Deblurgan-v2: Deblurring (orders-of-magnitude)
vibrant images even in less-than-ideal shooting conditions. faster and better. In Proceedings of the IEEE/CVF inter-
national conference on computer vision, pages 8878–8887,
6. Conclusion 2019. 1, 4
[6] Tzuu-Hseng S Li, Ching-Chang Chen, and Yu-Te Su. Opti-
In this project, we focus on GANs to address the chal- cal image stabilizing system using fuzzy sliding-mode con-
lenge of motion blur in images. By utilizing a GAN-based troller for digital cameras. IEEE Transactions on Consumer
model implemented in Keras, we focused on training and Electronics, 58(2):237–245, 2012. 4
evaluating our system using the GoPro dataset, which con- [7] Seungjun Nah, Tae Hyun Kim, and Kyoung Mu Lee. Deep
tains paired street view images that include both clear and multi-scale convolutional neural network for dynamic scene
deblurring. In Proceedings of the IEEE conference on
computer vision and pattern recognition, pages 3883–3891,
2017. 4
[8] Olaf Ronneberger, Philipp Fischer, and Thomas Brox. U-
net: Convolutional networks for biomedical image segmen-
tation. In Medical image computing and computer-assisted
intervention–MICCAI 2015: 18th international conference,
Munich, Germany, October 5-9, 2015, proceedings, part III
18, pages 234–241. Springer, 2015. 4
[9] Karen Simonyan. Very deep convolutional networks
for large-scale image recognition. arXiv preprint
arXiv:1409.1556, 2014. 2
[10] Shuochen Su, Mauricio Delbracio, Jue Wang, Guillermo
Sapiro, Wolfgang Heidrich, and Oliver Wang. Deep video
deblurring for hand-held cameras. In Proceedings of the
IEEE conference on computer vision and pattern recogni-
tion, pages 1279–1288, 2017. 4
[11] Linh Duy Tran, Son Minh Nguyen, and Masayuki Arai. Gan-
based noise model for denoising real images. In Proceedings
of the Asian Conference on Computer Vision, 2020. 1