Abstract
Many institutions, such as banks, usually require their customers to provide face images under proper illumination conditions. For some remote systems, a method that can automatically and objectively evaluate the illumination quality of a face image in a human-like manner is highly desired. However, few studies have been conducted in this area. To fill this research gap to some extent, we make two contributions in this paper. Firstly, in order to facilitate the study of illumination quality prediction for face images, a large-scale database, namely, Face Image Illumination Quality Database (FIIQD), is established. FIIQD contains 224,733 face images with various illumination patterns and for each image there is an associated illumination quality score. Secondly, based on deep convolutional neural networks (DCNN), a novel highly accurate model for predicting the illumination quality of face images is proposed. To make our results reproducible, the database and the source codes have been made publicly available at https://github.com/zhanglijun95/FIIQA.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Xia, D., Cui, D., Wang, J., Wang, Y.: A novel data schema integration framework for the human-centric services in smart city. ZTE Commun. 13(4), 25–33 (2015)
Sellahewa, H., Jassim, S.A.: Image-quality-based adaptive face recognition. IEEE Trans. IM 59(4), 805–813 (2010)
Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error visibility to structural similarity. IEEE Trans. IP 13(4), 600–612 (2004)
Truong, Q.C., Dang, T.K., Ha, T.: Face quality measure for face authentication. In: Dang, T.K., Wagner, R., Küng, J., Thoai, N., Takizawa, M., Neuhold, E. (eds.) FDSE 2016. LNCS, vol. 10018, pp. 189–198. Springer, Cham (2016). doi:10.1007/978-3-319-48057-2_13
Wong, Y., Chen, S., Mau, S., Sanderson, C., Lovell, B.C.: Patch-based probabilistic image quality assessment for face selection and improved video-based face recognition. In: IEEE CVPR Workshops, pp. 74–81 (2011)
Chen, J., Deng, Y., Bai, G., Su, G.: Face image quality assessment based on learning to rank. IEEE SPL 22(1), 90–94 (2015)
Mittal, A., Moorthy, A.K., Bovik, A.C.: No-reference image quality assessment in the spatial domain. IEEE Trans. IP 21(12), 4695–4708 (2012)
Mittal, A., Soundararajan, R., Bovik, A.C.: Making a ‘completely blind’ image quality analyzer. IEEE SPL 20(3), 209–212 (2013)
Liu, L., Liu, B., Huang, H., Bovik, A.C.: No-reference image quality assessment based on spatial and spectral entropies. Sig. Process. Image Commun. 29(8), 856–863 (2014)
Wu, Q., Wang, Z., Li, H.: A highly efficient method for blind image quality assessment. In: IEEE ICIP, pp. 339–343 (2015)
Zhang, L., Zhang, L., Bovik, A.C.: A feature-enriched completely blind image quality evaluator. IEEE Trans. IP 24(8), 2579–2591 (2015)
Wu, Q., Li, H., Meng, F., Ngan, K.N., Luo, B., Huang, C., Zeng, B.: Blind image quality assessment based on multichannel feature fusion and label transfer. IEEE Trans. CSVT 26(3), 425–440 (2016)
Liu, L., Hua, Y., Zhao, Q., Huang, H., Bovik, A.C.: Blind image quality assessment by relative gradient statistics and adaboosting neural network. Sig. Process. Image Commun. 40, 1–15 (2016)
Krizhevsky, A., Sutskever, I., Hinton, G.E.: ImageNet classification with deep convolutional neural networks. In: NIPS, pp. 1097–1105 (2012)
He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: IEEE CVPR, pp. 770–778 (2016)
Assembly, ITU Radiocommunication: Methodology for the Subjective Assessment of the Quality of Television Pictures. International Telecommunication Union (2003)
Georghiades, A.S., Belhumeur, P.N., Kriegman, D.J.: From few to many: Illumination cone models for face recognition under variable lighting and pose. IEEE Trans. PAMI 23(6), 643–660 (2001)
Sim, T., Baker, S., Bsat, M.: The CMU pose, illumination and expression database. IEEE Trans. PAMI 25(12), 1615–1618 (2003)
Phillips, P.J., Moon, H., Rizvi, S.A., Rauss, P.J.: The FERET evaluation methodology for face-recognition algorithms. IEEE Trans. PAMI 22(10), 1090–1104 (2000)
Chen, X., Chen, M., Jin, X., Zhao, Q.: Face illumination transfer through edge-preserving filters. In: IEEE CVPR, pp. 281–287 (2011)
Acknowledgments
This work was supported in part by the Natural Science Foundation of China under grant no. 61672380 and in part by the ZTE Industry-Academia-Research Cooperation Funds under grant no. CON1608310007.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2017 Springer International Publishing AG
About this paper
Cite this paper
Zhang, L., Zhang, L., Li, L. (2017). Illumination Quality Assessment for Face Images: A Benchmark and a Convolutional Neural Networks Based Model. In: Liu, D., Xie, S., Li, Y., Zhao, D., El-Alfy, ES. (eds) Neural Information Processing. ICONIP 2017. Lecture Notes in Computer Science(), vol 10636. Springer, Cham. https://doi.org/10.1007/978-3-319-70090-8_59
Download citation
DOI: https://doi.org/10.1007/978-3-319-70090-8_59
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-70089-2
Online ISBN: 978-3-319-70090-8
eBook Packages: Computer ScienceComputer Science (R0)