Abstract
The heterogeneous distributions of pixel intensities between natural scene and document images casts challenges for generalizing quality assessment models across these two types of images, where human perceptual scores and optical character recognition accuracy are the respective quality metrics. In this paper we propose a novel contractive generative adversarial model to learn a unified quality-aware representation of images from heterogeneous sources in a latent domain. We then build a unified image quality assessment framework by applying a regressor in the unveiled latent domain, where the regressor operates as if it is assessing the quality of a single type of images. Test results on blur distortion across three benchmarking datasets show that the proposed model achieves promising performance competitive to the state-of-the-art simultaneously for natural scene and document images.
This research is supported by the Auditing Digitisation Outputs in the Cultural Heritage Sector (ADOCHS) project (Contract No. BR/154/A6/ADOCHS), financed by the Belgian Science Policy (Belspo) within the scope of the BRAIN programme and by funding from the Flemish Government under the “Onderzoeksprogramma Artificiële Intelligentie (AI) Vlaanderen” programme.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Antonacopoulos, A., Downton, A.C.: Special issue on the analysis of historical documents. Int. J. Doc. Anal. Recogn. 9, 75–77 (2007)
Bosse, S., Maniry, D., Müller, K.R., Wiegand, T., Samek, W.: Deep neural networks for no-reference and full-reference image quality assessment. IEEE Trans. Image Process. 27, 206–219 (2018)
Cai, H., Li, L., Yi, Z., Gang, M.: Towards a blind image quality evaluator using multi-scale second-order statistics. Signal Process. Image Commun. 71, 88–99 (2019)
Goodfellow, I.J., et al.: Generative adversarial nets. In: Proceedings 27th International Conference on Neural Information Processing Systems, pp. 2672–2680 (2014)
Kang, L., Ye, P., Li, Y., Doermann, D.: Convolutional neural networks for no-reference image quality assessment. In: Proceedings 27th IEEE Conference on Computer Vision and Pattern Recognition, pp. 1733–1740 (2014)
Kang, L., Ye, P., Li, Y., Doermann, D.: A deep learning approach to document image quality assessment. In: Proceedings 2014 IEEE International Conference on Image Processing (ICIP), pp. 2570–2574 (2014)
Kim, J., Nguyen, A.D., Lee, S.: Deep cnn-based blind image quality predictor. IEEE Trans. Neural Netw. Learn. Syst. 30, 11–24 (2019)
Kim, J., Zeng, H., Ghadiyaram, D., Lee, S., Zhang, L., Bovik, A.C.: Deep convolutional neural models for picture-quality prediction challenges and solutions to data-driven image quality assessment. IEEE Signal Process. Mag. 34, 130–141 (2017)
Kumar, J., Ye, P., Doermann, D.: A dataset for quality assessment of camera captured document images. In: Iwamura, M., Shafait, F. (eds.) CBDAR 2013. LNCS, vol. 8357, pp. 113–125. Springer, Cham (2014). https://doi.org/10.1007/978-3-319-05167-3_9
Larson, E.C., Chandler, D.M.: Most apparent distortion: full-reference image quality assessment and the role of strategy. J. Electron. Imaging 19, 011006-1–011006-21 (2010)
Li, P., Peng, L., Cai, J., Ding, X., Ge, S.: Attention based rnn model for document image quality assessment. In: Proceedings 2017 14th IAPR International Conference on Document Analysis and Recognition (ICDAR), pp. 819–825 (2017)
Li, Q., Lin, W., Xu, J., Fang, Y.: Blind image quality assessment using statistical structural and luminance features. IEEE Trans. Multimedia 21, 3339–3352 (2013)
Li, Y., Po, L.M., Feng, L., Yuan, F.: No-reference image quality assessment with deep convolutional neural networks. In: Proceedings IEEE International Conference on Digital Signal Processing, pp. 685–689 (2016)
Lin, K.Y., Wang, G.: Hallucinated-iqa: no-reference image quality assessment via adversarial learning. In: Proceedings IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 732–741 (2018)
Liu, L., Liu, B., Huang, H., Bovik, A.C.: No-reference image quality assessment based on spatial and spectral entropies. Signal Process. Image Commun. 29, 856–863 (2014)
Lu, T., Dooms, A.: A deep transfer learning approach to document image quality assessment. In: Proceedings International Conference on Document Analysis and Recognition (ICDAR), pp. 1372–1377 (2019)
Lu, T., Dooms, A.: Towards content independent no-reference image quality assessment using deep learning. In: Proceedings IEEE 4th International Conference on Image, Vision and Computing (ICIVC), pp. 276–280 (2019)
Peng, X., Cao, H., Natarajan, P.: Document image ocr accuracy prediction via latent dirichlet allocation. In: Proceedings 13th International Conference on Document Analysis and Recognition (ICDAR), pp. 771–775 (2015)
Peng, X., Cao, H., Natarajan, P.: Document image quality assessment using discriminative sparse representation. In: Proceedings 12th IAPR Workshop on Document Analysis Systems (DAS), pp. 227–232 (2016)
Saad, M.A., Bovik, A.C., Charrier, C.: Blind image quality assessment - a natural scene statistics approach in the dct domain. IEEE Trans. Image Process. 21, 3339–3352 (2013)
Sheikh, H.R., Wang, Z., Cormack, L., Bovik, A.C.: Live image quality assessment database release 2, http://live.ece.utexas.edu/research/quality
Ye, P., Kumar, J., Kang, L., Doermann, D.: Unsupervised feature learning framework for no-reference image quality assessment. In: Proceedings IEEE Conference on Computer Vision and Pattern Recognition, pp. 1098–1105 (2012)
Zhu, J.Y., Park, T., Isola, P., Efros, A.A.: Unpaired image-to-image translation using cycle-consistent adversarial networks. In: Proceedings IEEE International Conference on Computer Vision (ICCV), pp. 2242–2251 (2017)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Lu, T., Dooms, A. (2020). A Novel Contractive GAN Model for a Unified Approach Towards Blind Quality Assessment of Images from Heterogeneous Sources. In: Bebis, G., et al. Advances in Visual Computing. ISVC 2020. Lecture Notes in Computer Science(), vol 12509. Springer, Cham. https://doi.org/10.1007/978-3-030-64556-4_3
Download citation
DOI: https://doi.org/10.1007/978-3-030-64556-4_3
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-64555-7
Online ISBN: 978-3-030-64556-4
eBook Packages: Computer ScienceComputer Science (R0)