条件生成对抗网络的低照度图像增强方法
Low-illumination image enhancement using a conditional generative adversarial network
- 2019年24卷第12期 页码:2149-2158
收稿:2019-04-18,
修回:2019-6-18,
录用:2019-6-25,
纸质出版:2019-12-16
DOI: 10.11834/jig.190145
移动端阅览

浏览全部资源
扫码关注微信
收稿:2019-04-18,
修回:2019-6-18,
录用:2019-6-25,
纸质出版:2019-12-16
移动端阅览
目的
2
在日常的图像采集工作中,由于场景光照条件差或设备的补光能力不足,容易产生低照度图像。为了解决低照度图像视觉感受差、信噪比低和使用价值低(难以分辨图像内容)等问题,本文提出一种基于条件生成对抗网络的低照度图像增强方法。
方法
2
本文设计一个具备编解码功能的卷积神经网络(CNN)模型作为生成模型,同时加入具备二分类功能的CNN作为判别模型,组成生成对抗网络。在模型训练的过程中,以真实的亮图像为条件,依靠判别模型监督生成模型以及结合判别模型与生成模型间的相互博弈,使得本文网络模型具备更好的低照度图像增强能力。在本文方法使用过程中,无需人工调节参数,图像输入模型后端到端处理并输出结果。
结果
2
将本文方法与现有方法进行比较,利用本文方法增强的图像在亮度、清晰度以及颜色还原度等方面有了较大的提升。在峰值信噪比、直方图相似度和结构相似性等图像质量评价指标方面,本文方法比其他方法的最优值分别提高了0.7 dB、3.9%和8.2%。在处理时间上,本文方法处理图像的速度远远超过现有的传统方法,可达到实时增强的要求。
结论
2
通过实验比较了本文方法与现有方法对于低照度图像的处理效果,表明本文方法具有更优的处理效果,同时具有更快的处理速度。
Objective
2
Low-illumination images are easily produced when taking pictures because of weak lighting conditions or devices with poor filling flash. Low-illumination images are difficult to recognize. Thus
the quality of low-illumination images needs to be improved. In the past
low-illumination image enhancement was dominated by histogram equalization (HE) and Retinex
but these methods cannot easily generate the desired results. Their results often entail problems
such as color distortion and blurred edges. A conditional generative adversarial network (CGAN)-based method is proposed to solve this poor visual perception problem. CGAN is an extension of the generative adversarial network (GAN). At present
it is widely used in data generation
including image de-raining
image resolution enhancement
and speech denoising. Unlike traditional low-illumination image enhancement methods that work on single image adjustment
this method achieves data-driven enhancement.
Method
2
This study proposes an encode-decode convolutional neural network (CNN) model as the generative model and a CNN model with a classification function as the discriminative model. The two models constitute a GAN. The model processes input images end to end and without adjusting the parameters manually. Instead of using synthetic image datasets
real-shot low-illumination images from the multi-exposure image dataset are used for training and testing. This image dataset contains multi-exposure sequential images
including under-and over-exposure images. The exposure of images is shifted by the exposure value (EV) of cameras or phones. Moreover
this dataset offers high-quality reference light images. During training
by offering reference light images from datasets as conditions to GAN
both models optimize their parameters according to the light images. As a result
the entire model is transformed into CGAN. The coding path of the generative model samples low-illumination images and processes the images at different scales. After coding
the encoding path restores the image size and shortens the distance between the outputs and conditional light images. The low-illumination images are denoised and restored by a different convolution processing of the generative model
and enhanced images are obtained. The discriminative model distinguishes the enhanced and reference light images by comparing their differences. The enhanced images are regarded as false
and the reference light images are regarded as true. Then
the discriminative model provides the result to the generative model. According to the feedback
the generative model optimizes the parameters to obtain an improved enhancement capability
and the discriminative model obtains an improved distinguishing capability by optimizing its own parameters. After training thousands of pairs of images
the parameters of both models are optimized. By using the discriminative model to supervise the generative model and by combining the interrelation between the two models
an improved image enhancement effect is achieved. When the proposed model is used to enhance low-illumination images
the discriminative model is no longer involved in the work
and the result is obtained directly from the generative model. Furthermore
skip connection and batch normalization are integrated into the proposed model. Skip connection transmits the gradient from shallow to deep layers. It has a transitional effect on the shallow and deep features. Batch normalization can effectively avoid gradient vanishing and explosion. Both approaches enhance the processing capability of the model.
Result
2
In this study
the entire network model and the single generative model are compared; the two sets of models represent CGAN and CNN methods
respectively.
Results
2
show that the entire network model has a better processing effect than the single generative model. This finding proves that the discriminative model improves the effect of the generative model during training. Afterward
eight existing methods are applied for comparison with the proposed method. By subjectively comparing the results of these methods
we find that our method achieves a better effect in terms of brightness
clarity
and color restoration. By using the peak signal-to-noise ratio (PSNR)
histogram similarity (HS)
and structural similarity (SSIM) as the objectives of comparison
our method exhibits improvements of 0.7 dB
3.9%
and 8.2%
respectively. Meanwhile
the processing time of each method is compared. By using a graphics processing unit (GPU) for acceleration
the proposed method becomes much faster than the other methods
especially traditional central processing unit (CPU)-based methods. The proposed method can meet the requirement of real-time applications. Furthermore
for several low-illumination images with bright parts
our method does not enhance these parts
whereas other existing methods always over-enhance the bright parts.
Conclusion
2
A conditional generative adversarial network-based method for low-illumination image enhancement is proposed. Experimental results show that the method proposed is more effective than existing methods not only in perception but also in speed.
Abdullah-Al-Wadud M, Kabir M H, Dewan M A A and Chae O. 2007. A dynamic histogram equalization for image contrast enhancement. IEEE Transactions on Consumer Electronics, 53(2):593-600[DOI:10.1109/TCE.2007.381734]
Cai J R, Gu S H and Zhang L. 2018. Learning a deep single image contrast enhancer from multi-exposure images. IEEE Transactions on Image Processing, 27(4):2049-2062[DOI:10.1109/TIP.2018.2794218]
Cheng Y, Deng D X, Yan J and Fan C E. 2019. Weakly illuminated image enhancement algorithm based on convolutional neural network. Journal of Computer Applications, 38(4):1162-1169
程宇, 邓德祥, 颜佳, 范赐恩. 2019.基于卷积神经网络的弱光照图像增强算法.计算机应用, 38(4):1162-1169)[DOI:10.11772/j.issn.1001-9081.2018091979]
Goodfellow I J, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, Courville A and Bengio Y. 2014. Generative adversarial nets//Proceedings of the 27th International Conference on Neural Information Processing Systems. Montreal, Canada: ACM, 2672-2680
Guo X J, Li Y and Ling H B. 2017. LIME:low-light image enhancement via illumination map estimation. IEEE Transactions on Image Processing, 26(2):982-993[DOI:10.1109/TIP.2016.2639450]
Isola P, Zhu J Y, Zhou T H and Efros A A. 2017. Image-to-image translation with conditional adversarial networks//Proceedings of 2017 IEEE Conference on Computer Vision and Pattern Recognition. Honolulu, HI, USA: IEEE, 5967-5976[ DOI:10.1109/CVPR.2017.632 http://dx.doi.org/10.1109/CVPR.2017.632 ]
Jobson D J, Rahman Z U and Woodell G A. 1997. A mulfiscale refinex for bridging the gop between color images and the human obseruation of scenes. IEEE Transactions onImage Processing A Publication of the IEEE Signal Processing Society, 6(7):965-976[DOI:10.1109/83.597272]
Ledig C, Theis L, Huszár F, Caballero J, Cunningham A, Acosta A, Aitken A, Tejani A, Totz J, Wang Z H and Shi W Z. 2017. Photo-realistic single image super-resolution using a generative adversarial network//Proceedings of 2017 IEEE Conference on Computer Vision and Pattern Recognition. Honolulu, HI, USA: IEEE, 105-114[ DOI:10.1109/CVPR.2017.19 http://dx.doi.org/10.1109/CVPR.2017.19 ]
Li C Y, Guo J C, Porikli F and Pang Y W. 2018. LightenNet:a convolutional neural network for weakly illuminated image enhancement. Pattern Recognition Letters, 104:15-22[DOI:10.1016/j.patrec.2018.01.010]
Lore K G, Akintayo A and Sarkar S. 2017. LLNet:a deep auto encoder approach to natural low-light image enhancement. Pattern Recognition, 61:650-662[DOI:10.1016/j.patcog.2016.06.008]
Li Q H, Bi D Y, Ma S P and He Y B. 2010. Image enhancement based on Retinex and vision adaptability. Journal of Image and Graphics, 15(12):1728-1732
李权合, 毕笃彦, 马时平, 何宜宝. 2010.基于Retinex和视觉适应性的图像增强.中国图象图形学报, 15(12):1728-1732)[DOI:10.11834/jig.20101221]
Ma H Q, Ma S P, Xu Y L and Zhu M M. 2019. Low-light image enhancement based on deep convolutional neural network. Acta Optica Sinica, 39(2):0210004
马红强, 马时平, 许悦雷, 朱明明. 2019.基于深度卷积神经网络的低照度图像增强.光学学报, 39(2):0210004
Mirza M and Osindero S. 2014. Conditional generative adversarial nets[EB/OL] .[2014-11-06]. https://arxiv.org/pdf/1411.1784.pdf https://arxiv.org/pdf/1411.1784.pdf
Radford A, Metz L and Chintala S. 2015. Unsupervised representation learning with deep convolutional generative adversarial networks[EB/OL][2019-04-03] . https://arxiv.org/pdf/1511.06434v1.pdf https://arxiv.org/pdf/1511.06434v1.pdf
Ren X T, Li M D, Cheng W H and Liu J Y. 2018. Joint enhancement and denoising method via sequential decomposition//Proceedings of 2018 IEEE International Symposium on Circuits and Systems. Florence, Italy: IEEE, 1-5[ DOI:10.1109/ISCAS.2018.8351427 http://dx.doi.org/10.1109/ISCAS.2018.8351427 ]
Ronneberger O, Fischer P and Brox T. 2015. U-net: convolutional networks for biomedical image segmentation//Proceedings of the 18th International Conference on Medical Image Computing and Computer-Assisted Intervention. Munich, Germany: Springer, 234-241[ DOI:10.1007/978-3-319-24574-4_28 http://dx.doi.org/10.1007/978-3-319-24574-4_28 ]
Tao L, Zhu C, Xiang G Q, Li Y, Jia H Z and Xie X D. 2017. LLCNN: a convolutional neural network for low-light image enhancement//Proceedings of 2017 IEEE Visual Communications and Image Processing. St. Petersburg, FL, USA: IEEE, 1-4[ DOI:10.1109/VCIP.2017.8305143 http://dx.doi.org/10.1109/VCIP.2017.8305143 ]
Wang Y F, Han J G and Fan L H. 2019. Algorithm research of speech enhancement based on WGAN. Journal of Chongqing University of Posts and Telecommunications:Natural Science Edition, 31(1):136-142
王怡斐, 韩俊刚, 樊良辉. 2019.基于WGAN的语音增强算法研究.重庆邮电大学学报:自然科学版, 31(1):136-142)[DOI:10.3979/j.issn.1673-825X.2019.01.018]
Ying Z Q, Li G, Ren Y R, Wang R G and Wang W M. 2017. A new image contrast enhancement algorithm using exposure fusion framework//Proceedings of the 17th International Conference on Computer Analysis of Images and Patterns. Ystad, Sweden: Springer, 36-46[ DOI:10.1007/978-3-319-64698-5_4 http://dx.doi.org/10.1007/978-3-319-64698-5_4 ]
Zhang H, Sindagi V and Patel V M. 2017. Image de-raining using a conditional generative adversarial network[EB/OL].[2017-02-04] . https://arxiv.org/pdf/1701.05957.pdf https://arxiv.org/pdf/1701.05957.pdf
相关作者
相关机构
京公网安备11010802024621