Blur Detection via Phase Spectrum

  • Renyan ZhangEmail author
Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 11366)


The effectiveness of blur features is very important in blur detection from a single image and the most existing blur features are sensitive to the strong edges in the blurred image region which degrades the detection methods. We analyze the information carried by the reconstruction of an image from the phase spectrum alone (RIPS) and the influences of blurring on RIPS. We find that a clear image region has more intensity changes than a blurred one because the former has more high frequency components. And the local maxima of RIPS are at where these image components occur, which make the RIPS of the clear image regions are obviously bigger than that of the blurred ones. Based on this finding, we proposed a simple blur feature, called Phase Map (PM), generated by thresholding RIPS adaptively. And our blur detection method propagates PM to the final blur map only by filtering PM using the relative total variation (RTV) filter. Our proposed method is evaluated on challenging blur image datasets. The evaluation demonstrates that PM feature is effective for different blur types and our detection method performs better than the state-of-the-art algorithms quantitatively and qualitatively.


Blur detection Phase spectrum 

Supplementary material

484523_1_En_46_MOESM1_ESM.pdf (177 kb)
Supplementary material 1 (pdf 176 KB)


  1. 1.
    Bae, S., Durand, F.: Defocus magnification. Comput. Graph. Forum 26(3), 571–579 (2007)CrossRefGoogle Scholar
  2. 2.
    Castleman, K.R.: Digital Image Process. Prentice Hall Press, New Jersey (1996)Google Scholar
  3. 3.
    Chakrabarti, A., Zickler, T., Freeman, W.T.: Analyzing spatially-varying blur. In: 2010 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2512–2519. IEEE Press (2010)Google Scholar
  4. 4.
    Chou, Y.C., Fang, C.Y., Su, P.C.: Content-based cropping using visual saliency and blur detection. In: 10th International Conference on Ubi-Media Computing and Workshops (Ubi-Media), pp. 1–6. IEEE Press (2017)Google Scholar
  5. 5.
    Elder, J.H., Zucker, S.W.: Local scale control for edge detection and blur estimation. IEEE Trans. Pattern Anal. Mach. Intell. 20(7), 699–716 (1998)CrossRefGoogle Scholar
  6. 6.
    Favaro, P., Soatto, S.: A geometric approach to shape from defocus. IEEE Trans. Pattern Anal. Mach. Intell. 27(3), 406–417 (2005)CrossRefGoogle Scholar
  7. 7.
    Fergus, R., Singh, B., Hertzmann, A., Roweis, S.T., Freeman, W.T.: Removing camera shake from a single photograph. ACM Trans. Graph. 25(3), 787–794 (2006)CrossRefGoogle Scholar
  8. 8.
    Ferzli, R., Karam, L.J.: A no-reference objective image sharpness metric based on the notion of just noticeable blur (JNB). IEEE Trans. Image Process. 18(4), 717–728 (2009)MathSciNetCrossRefGoogle Scholar
  9. 9.
    Forsyth, D.A., Ponce, J.: Computer Vision: A Modern Approach. Pearson, London (2012)Google Scholar
  10. 10.
    Golestaneh, S.A., Karam, L.J.: Spatially-varying blur detection based on multiscale fused and sorted transform coefficients of gradient magnitudes. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 596–605. IEEE Press (2017)Google Scholar
  11. 11.
    Ha, H., Im, S., Park, J., Jeon, H.G., Kweon, I.S.: High-quality depth from uncalibrated small motion clip. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5413–5421. IEEE Press (2016)Google Scholar
  12. 12.
    Levin, A.: Blind motion deblurring using image statistics. In: 2006 Advances in Neural Information Processing Systems (NIPS), vol. 19, pp. 841–848. MIT Press (2007)Google Scholar
  13. 13.
    Levin, A., Fergus, R., Durand, F., Freeman, W.T.: Image and depth from a conventional camera with a coded aperture. ACM Trans. Graph. 26(3), 70 (2007)CrossRefGoogle Scholar
  14. 14.
    Levin, A., Lischinski, D., Weiss, Y.: Colorization using optimization. ACM Trans. Graph. 23(3), 689–694 (2004)CrossRefGoogle Scholar
  15. 15.
    Li, S., Kang, X., Hu, J., Yang, B.: Image matting for fusion of multi-focus images in dynamic scenes. Inf. Fusion 14(2), 147–162 (2013)CrossRefGoogle Scholar
  16. 16.
    Liu, R., Li, Z., Jia, J.: Image partial blur detection and classification. In: 2008 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1–8. IEEE Press (2008)Google Scholar
  17. 17.
    Narvekar, N.D., Karam, L.J.: A no-reference image blur metric based on the cumulative probability of blur detection (CPBD). IEEE Trans. Image Process. 20(9), 2678–2683 (2011)MathSciNetCrossRefGoogle Scholar
  18. 18.
    Pan, J., Sun, D., Pfister, H., Yang, M.H.: Blind image deblurring using dark channel prior. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1628–1636. IEEE Press (2016)Google Scholar
  19. 19.
    Shi, J., Xu, L., Jia, J.: Discriminative blur detection features. In: 2014 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2965–2972. IEEE Press (2014)Google Scholar
  20. 20.
    Shi, J., Xu, L., Jia, J.: Just noticeable defocus blur detection and estimation. In: 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 657–665. IEEE Press (2015)Google Scholar
  21. 21.
    Su, B., Lu, S., Tan, C.L.: Blurred image region detection and classification. In: 19th International Conference on Multimedia, pp. 1397–1400. ACM Press (2011)Google Scholar
  22. 22.
    Tang, C., Hou, C., Song, Z.: Defocus map estimation from a single image via spectrum contrast. Opt. Lett. 38(10), 1706–1708 (2013)CrossRefGoogle Scholar
  23. 23.
    Tang, C., Wu, J., Hou, Y., Wang, P., Li, W.: A spectral and spatial approach of coarse-to-fine blurred image region detection. IEEE Sig. Process. Lett. 23(11), 1652–1656 (2016)CrossRefGoogle Scholar
  24. 24.
    Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 13(4), 600–612 (2004)CrossRefGoogle Scholar
  25. 25.
    Xu, L., Yan, Q., Xia, Y., Jia, J.: Structure extraction from texture via relative total variation. ACM Trans. Graph. 31(6), 139 (2012)Google Scholar
  26. 26.
    Yi, X., Eramian, M.: LBP-based segmentation of defocus blur. IEEE Trans. Image Process. 25(4), 1626–1638 (2016)MathSciNetCrossRefGoogle Scholar
  27. 27.
    Yu, F., Gallup, D.: 3D reconstruction from accidental motion. In: 2014 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3986–3993. IEEE Press (2014)Google Scholar
  28. 28.
    Zhu, X., Cohen, S., Schiller, S., Milanfar, P.: Estimating spatially varying defocus blur from a single image. IEEE Trans. Image Process. 22(12), 4879–4891 (2013)MathSciNetCrossRefGoogle Scholar
  29. 29.
    Zhuo, S., Sim, T.: Defocus map estimation from a single image. Pattern Recogn. 44(9), 1852–1858 (2011)CrossRefGoogle Scholar

Copyright information

© Springer Nature Switzerland AG 2019

Authors and Affiliations

  1. 1.College of Electrical Engineering and AutomationShandong University of Science and TechnologyQingdaoChina

Personalised recommendations