Abstract
At present, the majority of techniques for assessing image quality are limited to extracting features from an image in a single space. This paper proposes a new dual-space multi-feature fusion based method for full-reference image quality assessment. This method involves simultaneously extracting features from both the YIQ and L*a*b* color spaces. First, we extract the luminance, slope, chroma, and gradient features in the spatial domain of the image to describe the salient differences in the image. Second, based on contrast sensitivity characteristics, we extract spatial frequency features in the spatial domain of the image to represent frequency differences in the image. Next, merge the features extracted in the dual space to construct a quality perception feature vector. Finally, the feature vector is input into the Random Forest model for regression prediction to obtain the predicted score of the image. Many experiments have been carried out on the four public datasets, and contrasted with other methods. The experimental confirm that the proposed method predicts image quality more accurately. The MATLAB source code and dataset of this paper will be published on GitHub, and the corresponding author can be contacted if necessary.
Similar content being viewed by others
Data availability
All data generated or analyses during this study are included in this published paper. The datasets used and/or analysed during the current study available from the corresponding author on reasonable request.
References
Yang, F., Yang, C., An, P., et al.: 360 video quality assessment based on saliency-guided viewport extraction. Multimedia Syst. 30, 89 (2024). https://doi.org/10.1007/s00530-024-01285-0
No-Reference Screen Content Image Quality Assessment With Unsupervised Domain Adaptation: B. Chen, H. Li, H. Fan and S. Wang. IEEE Trans. Image Process. 30, 5463–5476 (2021). https://doi.org/10.1109/TIP.2021.3084750
S. Aghapour Maleki and H. Ghassemian.: Spatial Quality Assessment of Pansharpened Images Based on Gray Level Co-Occurrence Matrix. 2022 International Conference on Machine Vision and Image Processing (MVIP), Ahvaz, Iran, Islamic Republic of, 1–6 (2022). https://doi.org/10.1109/MVIP53647.2022.9738763
Kim, D., Lee, S., Kim, C.: Contextual Information Based Quality Assessment for Contrast-Changed Images. IEEE Signal Process. Lett. 26(1), 109–113 (2019). https://doi.org/10.1109/LSP.2018.2882916
Seo, S., Ki, S., Kim, M.: A Novel Just-Noticeable-Difference-Based Saliency-Channel Attention Residual Network for Full-Reference Image Quality Predictions. IEEE Trans. Circuits Syst. Video Technol. 31(7), 2602–2616 (2021). https://doi.org/10.1109/TCSVT.2020.3030895
Sun, W., Liao, Q., Xue, J.-H., Zhou, F.: SPSIM: A Superpixel-Based Similarity Index for Full-Reference Image Quality Assessment. IEEE Trans. Image Process. 27(9), 4232–4244 (2018). https://doi.org/10.1109/TIP.2018.2837341
Yang, J., Zhang, W., Li, X., Zhou, T., Ou, B.: Full Reference Image Quality Assessment by Considering Intra-Block Structure and Inter-Block Texture. IEEE Access 8, 179702–179715 (2020). https://doi.org/10.1109/ACCESS.2020.3028282
Sim, K., Yang, J., Lu, W., Gao, X.: MaD-DLS: Mean and Deviation of Deep and Local Similarity for Image Quality Assessment. IEEE Trans. Multimedia 23, 4037–4048 (2021). https://doi.org/10.1109/TMM.2020.3037482
Kim, W., Nguyen, A.-D., Lee, S., Bovik, A.C.: Dynamic Receptive Field Generation for Full-Reference Image Quality Assessment. IEEE Trans. Image Process. 29, 4219–4231 (2020). https://doi.org/10.1109/TIP.2020.2968283
Zhou, Z., Li, J., Quan, Y., Xu, R.: Image Quality Assessment Using Kernel Sparse Coding. IEEE Trans. Multimedia 23, 1592–1604 (2021). https://doi.org/10.1109/TMM.2020.3001472
Ding, K., Ma, K., Wang, S., Simoncelli, E.P.: Image Quality Assessment: Unifying Structure and Texture Similarity. IEEE Trans. Pattern Anal. Mach. Intell. 44(5), 2567–2581 (2022). https://doi.org/10.1109/TPAMI.2020.3045810
Dong, W., Bie, H., Lu, L., Li, Y.: Image Quality Assessment by Considering Multiscale and Multidirectional Visibility Differences in Shearlet Domain. IEEE Access 7, 78715–78728 (2019). https://doi.org/10.1109/ACCESS.2019.2922011
Gu, K., Li, L., Lu, H., Min, X., Lin, W.: A Fast Reliable Image Quality Predictor by Fusing Micro- and Macro-Structures. IEEE Trans. Industr. Electron. 64(5), 3903–3912 (2017). https://doi.org/10.1109/TIE.2017.2652339
Yu, M., Tang, Z., Zhang, X., Zhong, B., et al.: Perceptual Hashing With Complementary Color Wavelet Transform and Compressed Sensing for Reduced-Reference Image Quality Assessment. IEEE Trans. Circuits Syst. Video Technol. 32(11), 7559–7574 (2022). https://doi.org/10.1109/TCSVT.2022.3190273
Zhu, W., et al.: Multi-Channel Decomposition in Tandem With Free-Energy Principle for Reduced-Reference Image Quality Assessment. IEEE Trans. Multimedia 21(9), 2334–2346 (2019). https://doi.org/10.1109/TMM.2019.2902484
Shen, L., Zhao, B., Pan, Z., Peng, B., Kwong, S., Lei, J.: Channel Recombination and Projection Network for Blind Image Quality Measurement. IEEE Trans. Instrum. Meas. 71, 1–12 (2022). https://doi.org/10.1109/TIM.2022.3190029
Jin, C., Zhao, X., Xiong, Q., et al.: Blind Image Quality Assessment for Multiple Distortion Image. Circuits Syst Signal Process 41, 5807–5826 (2022). https://doi.org/10.1007/s00034-022-02055-x
Yang, X., Li, F., Liu, H.: TTL-IQA: Transitive Transfer Learning Based No-Reference Image Quality Assessment. IEEE Trans. Multimedia 23, 4326–4340 (2021). https://doi.org/10.1109/TMM.2020.3040529
Liu, L., Wang, T., Huang, H.: Pre-Attention and Spatial Dependency Driven No-Reference Image Quality Assessment. IEEE Trans. Multimedia 21(9), 2305–2318 (2019). https://doi.org/10.1109/TMM.2019.2900941
Wen, Y., Chen, J., Sheng, B., et al.: Structure-Aware Motion Deblurring Using Multi-Adversarial Optimized CycleGAN. IEEE Trans. Image Process. 30, 6142–6155 (2021). https://doi.org/10.1109/TIP.2021.3092814
Yue, G., Wu, H., Jiang, Q., et al.: Perceptual Quality Assessment of Retouched Face Images. IEEE Trans. Multimedia (2023). https://doi.org/10.1109/TMM.2023.3338412
Cui, Y., Jiang, G., Yu, M.Y., et al.: Stitched Wide Field of View Light Field Image Quality Assessment: Benchmark Database and Objective Metric. IEEE Trans. Multimedia 26, 5092–5107 (2024). https://doi.org/10.1109/TMM.2023.3330096
Fang, Z., Cui, Y., Yu, M., et al.: Blind 3D-Synthesized Image Quality Measurement by Analysis of Local and Global Statistical Properties. IEEE Trans. Instrum. Meas. 72, 1–15 (2023). https://doi.org/10.1109/TIM.2023.3306527
Baig, M.A., Moinuddin, A.A., Khan, E.: A simple spatial domain method for quality evaluation of blurred images. Multimedia Syst. 30, 28 (2024). https://doi.org/10.1007/s00530-023-01223-6
Z. Zhang, W. Sun, X. Min, T. Wang, W. Lu and G. Zhai.: A Full-Reference Quality Assessment Metric for Fine-Grained Compressed Images. 2021 International Conference on Visual Communications and Image Processing (VCIP), Munich, Germany. 1–4 (2021). https://doi.org/10.1109/VCIP53242.2021.9675389
Liu, Y., Li, X.: No-Reference Quality Assessment for Contrast-Distorted Image. IEEE Access 8, 84105–84115 (2020). https://doi.org/10.1109/ACCESS.2020.2991842
Lin, X., Sun, S., Huang, W., et al.: EAPT: Efficient Attention Pyramid Transformer for Image Processing. IEEE Trans. Multimedia 25, 50–61 (2023). https://doi.org/10.1109/TMM.2021.3120873
Sheikh, H.R., Sabir, M.F., Bovik, A.C.: A Statistical Evaluation of Recent Full Reference Image Quality Assessment Algorithms. IEEE Trans. Image Process. 15(11), 3440–3451 (2006). https://doi.org/10.1109/TIP.2006.881959
Sheikh, H.R., Bovik, A.C.: Image information and visual quality. IEEE Trans. Image Process. 15(2), 430–444 (2006). https://doi.org/10.1109/TIP.2005.859378
Zhou, W., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 13(4), 600–612 (2004). https://doi.org/10.1109/TIP.2003.819861
Zhang, L., Zhang, L., Mou, X., Zhang, D.: FSIM: A Feature Similarity Index for Image Quality Assessment. IEEE Trans. Image Process. 20(8), 2378–2386 (2011). https://doi.org/10.1109/TIP.2011.2109730
Liu, W.L., Narwaria, M.: Image Quality Assessment Based on Gradient Similarity. IEEE Trans. Image Process. 21(4), 1500–2151 (2012). https://doi.org/10.1109/TIP.2011.2175935
Zhang, L., Shen, Y.: VSI A Visual Saliency-Induced Index for Perceptual Image Quality Assessment. IEEE Trans. Image Process. 23(10), 4270–4281 (2014). https://doi.org/10.1109/TIP.2014.2346028
Ma, Y., Hao, F., Liu, G., et al.: Full-Reference Color Image Quality Assessment Method Based on Extreme Learning Machine for Low-Level Features. Journal of Nanjing Normal University (Natural Science Edition) 45(4), 91–101 (2022). https://doi.org/10.3969/j.issn.1001-4616.2022.04.013
Zhou, W., Shi, L., Chen, Z., et al.: Tensor Oriented No-Reference Light Field Image Quality Assessment. IEEE Trans. Image Process. 29, 4070–4084 (2020). https://doi.org/10.1109/TIP.2020.2969777
Acknowledgements
This work is supported by Science and Technology Research Project of Jiangxi Provincial Department of Education (GJJ2200529).
Author information
Authors and Affiliations
Contributions
X.W wrote the main manuscript text, Z.S provide data support
Corresponding author
Ethics declarations
Conflicts of interests
There is no conflict in this paper.
Ethical approval
The main content of this paper is the study of IQA methods, and there is very close relevance between the topic of the revised manuscript and the scope of the Computer Graphics Society’s Journals.
Additional information
Communicated by Q. Shen.
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Wu, X., Shi, Z. Full reference image quality assessment based on dual-space multi-feature fusion. Multimedia Systems 30, 151 (2024). https://doi.org/10.1007/s00530-024-01353-5
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s00530-024-01353-5