The Visual Computer

, 24:281 | Cite as

Digital shallow depth-of-field adapter for photographs

  • Kyuman Jeong
  • Dongyeon Kim
  • Soon-Yong Park
  • Seungyong Lee
Original Article


This paper proposes a novel method to synthesize shallow depth-of-field images from two input photographs taken with different aperture values. The basic approach is to estimate the depth map of a given scene using a DFD (depth-from-defocus) algorithm and blur an input image according to the estimated depth map. The depth information estimated by DFD contains much noise and error, while the estimation is rather accurate along the edges of the image. To overcome the limitation, we propose a depth map filling algorithm using a set of initial depth maps and a segmented image. After depth map filling, the depth map can be fine tuned by applying segment clustering and user interaction. Since our method blurs an input image according to the estimated depth information, it generates physically plausible result images with shallow depth-of-field. In addition to depth-of-field control, the proposed method can be utilized for digital refocusing and detail control in image stylization.


Depth-of-field control Depth estimation Depth from defocus Digital refocusing Image detail control 


  1. 1.
    CombineZ5. Scholar
  2. 2.
    Depth-of-field adapter. Scholar
  3. 3.
    Depth of Field Generator Pro. Scholar
  4. 4.
    Helicon Focus. Scholar
  5. 5.
    Letus35. Scholar
  6. 6.
    SGPro. Scholar
  7. 7.
    Shallow depth of field tutorial. Scholar
  8. 8.
    Simulating shallow depth of field with the GIMP. Scholar
  9. 9.
    Agarwala, A., Dontcheva, M., Agrawala, M., Drucker, S., Colburn, A., Curless, B., Salesin, D., Cohen, M.: Interactive digital photomontage. ACM Trans. Graph. 23, 294–302 (2004)CrossRefGoogle Scholar
  10. 10.
    Cho, S., Tam, W.J., Speranza, F., Renaud, R., Hur, N., Lee, S.I.: Depth maps created from blur information using images with focus at near and at far. Proc. SPIE 6055, 462–473 (2006)Google Scholar
  11. 11.
    Christoudias, C., Georgescu, B., Meer, P.: Synergism in lowlevel vision. In: Proceedings of the 16th International Conference on Pattern Recognition, pp. 150–155. Quebec City, QC (2002)Google Scholar
  12. 12.
    Comaniciu, D., Meer, P.: Mean shift: a robust approach toward feature space analysis. IEEE Trans. Pattern Anal. Mach. Intell. 24, 603–619 (2002)CrossRefGoogle Scholar
  13. 13.
    Cosgrove, R.: Creating depth of field. Digit. Photogr. Buyer User, p. 89 (2003)Google Scholar
  14. 14.
    Forster, B., Ville, D.V.D., Berent, J., Sage, D., Unser, M.: Complex wavelets for extended depth-of-field: a new method for the fusion of multichannel microscopy images. Microsc. Res. Tech. 65, 33–42 (2004)CrossRefGoogle Scholar
  15. 15.
    Forster, B., Ville, D.V.D., Berent, J., Sage, D., Unser, M.: Extended depth-of-focus for multi-channel microscopy images: a complex wavelet approach. In: Proceedings of the Second 2004 IEEE International Symposium on Biomedical Imaging: From Nano to Macro, pp. 660–663. Arlington, VA (2004)Google Scholar
  16. 16.
    Lacey, R.: Creating depth of field in Paint Shop Pro.∼rldepthoffield.htmGoogle Scholar
  17. 17.
    Lin, J., Zhang, C., Shi, Q.: Estimating the amount of defocus through a wavelet transform approach. Pattern Recogn. Lett. 25, 407–411 (2004)CrossRefGoogle Scholar
  18. 18.
    London, B., Upton, J., Kobre, K., Brill, B.: Photography, 8th edn. Prentice Hall (2004)Google Scholar
  19. 19.
    Merklinger, H.M.: The ins and outs of focus. Internet edition is available in PDF at Scholar
  20. 20.
    Ng, R.: Fourier slice photography. ACM Trans. Graph. 24, 735–744 (2005)CrossRefGoogle Scholar
  21. 21.
    Ng, R., Levoy, M., Bredif, M., Duval, G., Horowitz, M., Hanrahan, P.: Light field photography with a hand-held plenoptic camera. Stanford University Computer Science Tech. Rep. (2005)Google Scholar
  22. 22.
    Pentland, A.P.: A new sense for depth of field. IEEE Trans. Pattern Anal. Mach. Intell. 9, 523–531 (1987)CrossRefGoogle Scholar
  23. 23.
    Raskar, R., Tumblin, J.: SIGGRAPH ’2005 course notes on computational photography, SIGGRAPH. ACM (2005)Google Scholar
  24. 24.
    Raskar, R., Tumblin, J., Levoy, M., Nayer, S.: SIGGRAPH ’2006 course notes on computational photography, SIGGRAPH. ACM (2006)Google Scholar
  25. 25.
    Ray, S.F.: Applied Photographic Optics, 3rd edn. Focal Press (1988)Google Scholar
  26. 26.
    Subbarao, M., Surya, G.: Depth from defocus: a spatial domain approach. Int. J. Comput. Vis. 13, 271–294 (1994)CrossRefGoogle Scholar
  27. 27.
    Surya, G.: Three-dimensional scene recovery from image defocus. Ph.D. thesis, Department of Electrical Engineering, State University of New York (SUNY) at Stony Brook (1994)Google Scholar
  28. 28.
    Tucker, S.C., Cathey, W.T., Dowski, E.R.: Extended depth of field and aberration control for inexpensive digital microscope systems. Opt. Express 4, 467 (1999)CrossRefGoogle Scholar
  29. 29.
    Watanabe, M., Nayar, S.K.: Rational filters for passive depth from defocus. Int. J. Comput. Vis. 27, 203–225 (1998)CrossRefGoogle Scholar
  30. 30.
    Weiss, B.: Fast median and bilateral filtering. ACM Trans. Graph. 25, 519–526 (2006)CrossRefGoogle Scholar
  31. 31.
    Winnemöller, H., Olsen, S.C., Gooch, B.: Real-time video abstraction. ACM Trans. Graph. 25, 1221–1226 (2006)CrossRefGoogle Scholar
  32. 32.
    Ziou, D., Deschenes, F.: Depth from defocus estimation in spatial domain. Comput. Vis. Image Understand. 81, 143–165 (2001)CrossRefMATHGoogle Scholar

Copyright information

© Springer-Verlag 2007

Authors and Affiliations

  • Kyuman Jeong
    • 2
  • Dongyeon Kim
    • 1
  • Soon-Yong Park
    • 3
  • Seungyong Lee
    • 1
  1. 1.Dept. of Computer Sci. &Eng.POSTECHPohangKorea
  2. 2.Mobile Communication DivisionSamsung Electronics Co. Ltd.SuwonKorea
  3. 3.Dept. of Computer EngineeringKyungpook National UniversityDaeguKorea

Personalised recommendations