Hierarchical Interpolation-Based Disocclusion Region Recovery for Two-View to N-View Conversion System

  • Wun-Ting Lin
  • Chen-Ting Yeh
  • Shang-Hong LaiEmail author
Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 9315)


In this paper, we propose a novel disocclusion region recovery approach for two-view to n-view conversion system. Although the topic of view synthesis has been exhaustively studied for decades, a reliable disocclusion region recovery approach, an indispensable issue in synthesizing realistic content of virtual view, is still under research. The most common concept used for predicting these unknown pixels is inpainting-related method, which fills the disocclusion region with the information of mated exemplars in self-defined searching domain. In spite of widely taken in making up the missing values generated among the synthesis procedures, the result quality of inpainting-based approach is sensitive to the filling priority and also unstable in recovering large disocclusion region. Therefore, we propose a hierarchical interpolation-based approach to calculate the desired lost information under coarse-to-fine manner accompanied with the joint bilateral upsampling technology, applied for enlarging the estimation from small dimension to higher-resolution. Proposed hierarchical interpolation-based scheme is more robust in restoring the value of missing region and also induces fewer artifacts. We demonstrate the superior quality of the synthesized virtual views under the proposed recovery algorithm over the traditional inpainting-based method through experiments on several benchmarking video datasets.


Disocclusion region recovery Image inpainting View interpolation Novel view synthesis 



This work was partially supported by Ministry of Science and Technology, Taiwan, R.O.C., under the grant MOST 101-2221-E-007-129-MY3.


  1. 1.
    Criminisi, A., Perez, P., Toyama, K.: Object removal by exemplar-based inpainting. In: Proceedings of 2003 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2003. vol. 2, pp. II-721. IEEE (2003)Google Scholar
  2. 2.
    Do, L., Zinger, S., et al.: Quality improving techniques for free-viewpoint DIBR. In: IS&T/SPIE Electronic Imaging. pp. 75240I–75240I. International Society for Optics and Photonics (2010)Google Scholar
  3. 3.
    Kopf, J., Cohen, M.F., Lischinski, D., Uyttendaele, M.: Joint bilateral upsampling. ACM Trans. Graph. (TOG). vol. 26, Article no. 96. ACM (2007)Google Scholar
  4. 4.
    Lin, S.-J., Cheng, C.-M., Lai, S.-H.: Spatio-temporally consistent multi-view video synthesis for autostereoscopic displays. In: Muneesawang, P., Wu, F., Kumazawa, I., Roeksabutr, A., Liao, M., Tang, X. (eds.) PCM 2009. LNCS, vol. 5879, pp. 532–542. Springer, Heidelberg (2009)CrossRefGoogle Scholar
  5. 5.
    Luo, H.L., Shen, C.T., Chen, Y.C., Wu, R.H., Hung, Y.P.: Automatic multi-resolution joint image smoothing for depth map refinement. In: 2013 2nd IAPR Asian Conference on Pattern Recognition (ACPR), pp. 284–287. IEEE (2013)Google Scholar
  6. 6.
    MPEG ISO/IEC JTC1/SC29/WG11: View Synthesis Software Manual (VSRS), release 3.5 (2009)Google Scholar
  7. 7.
    Oh, K.J., Yea, S., Ho, Y.S.: Hole filling method using depth based in-painting for view synthesis in free viewpoint television and 3D video. In: Picture Coding Symposium, 2009. PCS 2009. pp. 1–4. IEEE (2009)Google Scholar
  8. 8.
    Scharstein, D., Pal, C.: Learning conditional random fields for stereo. In: IEEE Conference on Computer Vision and Pattern Recognition, 2007. CVPR 2007. pp. 1–8. IEEE (2007)Google Scholar
  9. 9.
    Solh, M., AlRegib, G.: Hierarchical hole-filling for depth-based view synthesis in FTV and 3D video. IEEE J. Select. Topics Signal Process. 6(5), 495–504 (2012)CrossRefGoogle Scholar
  10. 10.
    Wang, L., Jin, H., Yang, R., Gong, M.: Stereoscopic inpainting: joint color and depth completion from stereo images. In: IEEE Conference on Computer Vision and Pattern Recognition, 2008, CVPR 2008. pp. 1–8. IEEE (2008)Google Scholar
  11. 11.
    Wei, C.-H., Chiang, C.-K., Sun, Y.-W., Lin, M.-H., Lai, S.-H.: Novel multi-view synthesis from a stereo image pair for 3D display on mobile phone. In: Park, J.-I., Kim, J. (eds.) ACCV 2012. LNCS, vol. 7729, pp. 568–579. Springer, Heidelberg (2013)CrossRefGoogle Scholar
  12. 12.
    Yang, Q.: A non-local cost aggregation method for stereo matching. In: 2012 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1402–1409. IEEE (2012)Google Scholar
  13. 13.
    Zhao, Y., Zhu, C., Chen, Z., Yu, L.: Depth no-synthesis-error model for view synthesis in 3D video. IEEE Trans. Image Process. 20(8), 2221–2228 (2011)MathSciNetCrossRefGoogle Scholar

Copyright information

© Springer International Publishing Switzerland 2015

Authors and Affiliations

  1. 1.Department of Computer ScienceNational Tsing Hua UniversityHsinchuTaiwan

Personalised recommendations