Skip to main content

Advertisement

SpringerLink
  1. Home
  2. Computational Visual Media
  3. Article
3D corrective nose reconstruction from a single image
Download PDF
Your article has downloaded

Similar articles being viewed by others

Slider with three articles shown per slide. Use the Previous and Next buttons to navigate the slides or the slide controller buttons at the end to navigate through each slide.

What Does 2D Geometric Information Really Tell Us About 3D Face Shape?

25 July 2019

Anil Bas & William A. P. Smith

Patch-based facial texture super-resolution by fitting 3D face models

13 February 2019

Chengchao Qu, Eduardo Monari, … Jürgen Beyerer

Holistic 3D face and head reconstruction with geometric details from a single image

16 August 2022

Jungwoo Lee, Jonathan Samuel Lumentut & In Kyu Park

3D face dense reconstruction based on sparse points using probabilistic principal component analysis

09 November 2021

Xiaoxiao Xie, Xingce Wang & Zhongke Wu

Reg3DFacePtCd: Registration of 3D Point Clouds Using a Common Set of Landmarks for Alignment of Human Face Images

08 May 2019

Parama Bagchi, Debotosh Bhattacharjee & Mita Nasipuri

Towards reduced dependency and faster unsupervised 3D face reconstruction

13 February 2023

Hitika Tiwari, Venkatesh K. Subramanian & Yong-Sheng Chen

Sign-correlation cascaded regression for face alignment

12 June 2019

Dansong Cheng, Yongqiang Zhang, … Xiaofang Liu

Joint 3D facial shape reconstruction and texture completion from a single image

06 December 2021

Xiaoxing Zeng, Zhelun Wu, … Yu Qiao

Personalized face-pose estimation network using incrementally updated face shape parameters

27 January 2022

Makoto Sei, Akira Utsumi, … Joo-Ho Lee

Download PDF
  • Research Article
  • Open Access
  • Published: 06 December 2021

3D corrective nose reconstruction from a single image

  • Yanlong Tang1,
  • Yun Zhang2,
  • Xiaoguang Han3,
  • Fang-Lue Zhang4,
  • Yu-Kun Lai5 &
  • …
  • Ruofeng Tong6 

Computational Visual Media volume 8, pages 225–237 (2022)Cite this article

  • 922 Accesses

  • 1 Citations

  • Metrics details

Abstract

There is a steadily growing range of applications that can benefit from facial reconstruction techniques, leading to an increasing demand for reconstruction of high-quality 3D face models. While it is an important expressive part of the human face, the nose has received less attention than other expressive regions in the face reconstruction literature. When applying existing reconstruction methods to facial images, the reconstructed nose models are often inconsistent with the desired shape and expression. In this paper, we propose a coarse-to-fine 3D nose reconstruction and correction pipeline to build a nose model from a single image, where 3D and 2D nose curve correspondences are adaptively updated and refined. We first correct the reconstruction result coarsely using constraints of 3D-2D sparse landmark correspondences, and then heuristically update a dense 3D-2D curve correspondence based on the coarsely corrected result. A final refinement step is performed to correct the shape based on the updated 3D-2D dense curve constraints. Experimental results show the advantages of our method for 3D nose reconstruction over existing methods.

Download to read the full article text

Working on a manuscript?

Avoid the common mistakes

References

  1. Samad, M. D.; Iftekharuddin, K. M. Frenet frame-based generalized space curve representation for pose-invariant classification and recognition of 3-D face. IEEE Transactions on Human-Machine Systems Vol. 46, No. 4, 522–533, 2016.

    Article  Google Scholar 

  2. Werghi, N.; Tortorici, C.; Berretti, S.; Del Bimbo, A. Boosting 3D LBP-based face recognition by fusing shape and texture descriptors on the mesh. IEEE Transactions on Information Forensics and Security Vol. 11, No. 5, 964–979, 2016.

    Article  Google Scholar 

  3. Li, H.; Yu, J. H.; Ye, Y. T.; Bregler, C. Realtime facial animation with on-the-fly correctives. ACM Transactions on Graphics Vol. 32, No. 4, Article No. 42, 2013.

    Google Scholar 

  4. Li, Y.; Ma, L. Q.; Fan, H. Q.; Mitchell, K. Feature-preserving detailed 3D face reconstruction from a single image. In: Proceedings of the 15th ACM SIGGRAPH European Conference on Visual Media Production, Article No. 1, 2018.

  5. Jiang, L.; Zhang, J. Y.; Deng, B. L.; Li, H.; Liu, L. G. 3D face reconstruction with geometry details from a single image. IEEE Transactions on Image Processing Vol. 27, No. 10, 4756–4770, 2018.

    Article  MathSciNet  Google Scholar 

  6. Tang, Y. L.; Han, X. G.; Li, Y.; Ma, L. Q.; Tong, R. F. Expressive facial style transfer for personalized memes mimic. The Visual Computer Vol. 35, 783–795, 2019.

    Article  Google Scholar 

  7. Kass, M.; Witkin, A.; Terzopoulos, D. Snakes: Active contour models. International Journal of Computer Vision Vol. 1, No. 4, 321–331, 1988.

    Article  Google Scholar 

  8. Tang, Y. L.; Tong, R. F.; Tang, M.; Zhang, Y. Depth incorporating with color improves salient object detection. The Visual Computer Vol. 32, 111–121, 2016.

    Article  Google Scholar 

  9. Blanz, V.; Vetter, T. A morphable model for the synthesis of 3D faces. In: Proceedings of the 26th Annual Conference on Computer Graphics and Interactive Techniques, 187–194, 1999.

  10. Paysan, P.; Knothe, R.; Amberg, B.; Romdhani, S.; Vetter, T. A 3D face model for pose and illumination invariant face recognition. In: Proceedings of the 6th IEEE International Conference on Advanced Video and Signal Based Surveillance, 296–301, 2009.

  11. Booth, J.; Roussos, A.; Zafeiriou, S.; Ponniah, A.; Dunaway, D. A 3D morphable model learnt from 10,000 faces. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 5543–5552, 2016.

  12. Zhu, X. Y.; Lei, Z.; Liu, X. M.; Shi, H. L.; Li, S. Z. Face alignment across large poses: A 3D solution. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 146–155, 2016.

  13. Cao, C.; Weng, Y. L.; Zhou, S.; Tong, Y. Y.; Zhou, K. FaceWarehouse: A 3D facial expression database for visual computing. IEEE Transactions on Visualization and Computer Graphics Vol. 20, No. 3, 413–425, 2014.

    Article  Google Scholar 

  14. Feng, Y.; Wu, F.; Shao, X. H.; Wang, Y. F.; Zhou, X. Joint 3D face reconstruction and dense alignment with position map regression network. In: Computer Vision-ECCV 2018. Lecture Notes in Computer Science, Vol.11218. Ferrari, V.; Hebert, M.; Sminchisescu, C.; Weiss, Y. Eds. Springer Cham, 557–574, 2018.

    Chapter  Google Scholar 

  15. Sanyal, S.; Bolkart, T.; Feng, H. W.; Black, M. J. Learning to regress 3D face shape and expression from an image without 3D supervision. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 7755–7764, 2019.

  16. Deng, Y.; Yang, J. L.; Xu, S. C.; Chen, D.; Jia, Y. D.; Tong, X. Accurate 3D face reconstruction with weakly-supervised learning: From single image to image set. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 285–295, 2019.

  17. Garrido, P.; Zollhöfer, M.; Casas, D.; Valgaerts, L.; Varanasi, K.; Pérez, P.; Theobalt, C. Reconstruction of personalized 3D face rigs from monocular video. ACM Transactions on Graphics Vol. 35, No. 3, Article No. 28, 2016.

    Google Scholar 

  18. Sorkine, O.; Cohen-Or, D.; Lipman, Y.; Alexa, M.; Rössl, C.; Seidel, H. P. Laplacian surface editing. In: Proceedings of the Eurographics/ACM SIGGRAPH symposium on Geometry processing, 175–184, 2004.

  19. Bouaziz, S.; Wang, Y. G.; Pauly, M. Online modeling for realtime facial animation. ACM Transactions on Graphics Vol. 32, No. 4, Article No. 40, 2013.

    Google Scholar 

  20. Vallet, B.; Lévy, B. Spectral geometry processing with manifold harmonics. Computer Graphics Forum Vol. 27, No. 2, 251–260, 2008.

    Article  Google Scholar 

  21. Wen, Q.; Xu, F.; Lu, M.; Yong, J. H. Real-time 3D eyelids tracking from semantic edges. ACM Transactions on Graphics Vol. 36, No. 6, Article No. 193, 2017.

    Google Scholar 

  22. Garrido, P.; Zollhöfer, M.; Wu, C. L.; Bradley, D.; Pérez, P.; Beeler, T.; Theobalt, C. Corrective 3D reconstruction of lips from monocular video. ACM Transactions on Graphics Vol. 35, No. 6, Article No. 219, 2016.

    Google Scholar 

  23. Dinev, D.; Beeler, T.; Bradley, D.; Bächer, M.; Xu, H.; Kavan, L. User-guided lip correction for facial performance capture. Computer Graphics Forum Vol. 37, No. 8, 93–101, 2018.

    Article  Google Scholar 

  24. Feng, Z. H.; Huber, P.; Kittler, J.; Hancock, P.; Wu, X. J.; Zhao, Q. J.; Koppen, P.; Raetsch, M. Evaluation of dense 3D reconstruction from 2D face images in the wild. In: Proceedings of the 13th IEEE International Conference on Automatic Face & Gesture Recognition, 780–786, 2018.

  25. Yin, L. J.; Wei, X. Z.; Sun, Y.; Wang, J.; Rosato, M. J. A 3D facial expression database for facial behavior research. In: Proceedings of the 7th International Conference on Automatic Face and Gesture Recognition, 211–216, 2006.

  26. Thies, J.; Zollhöfer, M.; Stamminger, M.; Theobalt, C.; Nießner, M. Face2Face: Real-time face capture and reenactment of RGB videos. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2387–2395, 2016.

  27. Canny, J. A computational approach to edge detection. IEEE Transactions on Pattern Analysis and Machine Intelligence Vol. PAMI-8, No. 6, 679–698, 1986.

    Article  Google Scholar 

Download references

Acknowledgements

This research was supported by the National Natural Science Foundation of China (Grant Nos. 61972342, 61602402, and 61902334), Zhejiang Provincial Basic Public Welfare Research (Grant No. LGG19F020001), Shenzhen Fundamental Research (General Project) (Grant No. JCYJ20190814112007258), and the Royal Society (Grant No. IES\R1\180126).

Author information

Authors and Affiliations

  1. Tencent Games Lightspeed & Quantum Studios, Shenzhen, China

    Yanlong Tang

  2. Communication University of Zhejiang, Hangzhou, China

    Yun Zhang

  3. Shenzhen Research Institute of Big Data, the Chinese University of Hong Kong (Shenzhen), Shenzhen, China

    Xiaoguang Han

  4. Victoria University of Wellington, Wellington, New Zealand

    Fang-Lue Zhang

  5. Cardiff University, Wales, UK

    Yu-Kun Lai

  6. Zhejiang University, Hangzhou, China

    Ruofeng Tong

Authors
  1. Yanlong Tang
    View author publications

    You can also search for this author in PubMed Google Scholar

  2. Yun Zhang
    View author publications

    You can also search for this author in PubMed Google Scholar

  3. Xiaoguang Han
    View author publications

    You can also search for this author in PubMed Google Scholar

  4. Fang-Lue Zhang
    View author publications

    You can also search for this author in PubMed Google Scholar

  5. Yu-Kun Lai
    View author publications

    You can also search for this author in PubMed Google Scholar

  6. Ruofeng Tong
    View author publications

    You can also search for this author in PubMed Google Scholar

Corresponding author

Correspondence to Yun Zhang.

Additional information

Yanlong Tang is currently a researcher at Tencent. He obtained his Ph.D. degree in 2019 from Zhejiang University, and his B.Sc. degree from Shandong University in 2013. His research interests include 3D face reconstruction, image processing, and computer vision.

Yun Zhang is an associate professor at Zhejiang Communication University. He received his doctoral degree from Zhejiang University in 2013, and bachelor and master degrees from Hangzhou Dianzi University in 2006 and 2009, respectively. In 2018, he was a visiting scholar at Cardiff University. His research interests include computer graphics, image and video editing, computer vision, and virtual reality. He is a member of the CCF.

Xiaoguang Han received his B.Sc. degree in mathematics in 2009 from Nanjing University of Aeronautics and Astronatics and his M.Sc. degree in applied mathematics in 2011 from Zhejiang University. He obtained his Ph.D. degree in 2017 from the University of Hong Kong. He is currently an assistant professor at Shenzhen Research Institute of Big Data, the Chinese University of Hong Kong (Shenzhen). His research mainly focuses on computer vision, computer graphics, and 3D deep learning.

Fang-Lue Zhang is currently a lecturer at Victoria University of Wellington, New Zealand. He received his bachelor degree from Zhejiang University in 2009, and his doctoral degree from Tsinghua University in 2015. His research interests include image and video editing, computer vision, and computer graphics. He is a member of the IEEE and ACM. He received a Victoria Early-Career Research Excellence Award in 2019 and a Marsden Fast-Start grant from the New Zealand Royal Society in 2021.

Yu-Kun Lai received his bachelor degree and Ph.D. degree in computer science from Tsinghua University in 2003 and 2008, respectively. He is currently a professor in the School of Computer Science & Informatics, Cardiff University. His research interests include computer graphics, geometry processing, image processing, and computer vision. He is on the editorial boards of Computer Graphics Forum and The Visual Computer.

Ruofeng Tong is a professor in the Department of Computer Science, Zhejiang University. He received his B.Sc. degree from Fudan University in 1991 and obtained his Ph.D. degree from Zhejiang University in 1996. His research interests include image and video processing, computer graphics, and computer animation.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduc-tion in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made.

The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder.

To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

Other papers from this open access journal are available free of charge from http://www.springer.com/journal/41095. To submit a manuscript, please go to https://www.editorialmanager.com/cvmj.

Reprints and Permissions

About this article

Verify currency and authenticity via CrossMark

Cite this article

Tang, Y., Zhang, Y., Han, X. et al. 3D corrective nose reconstruction from a single image. Comp. Visual Media 8, 225–237 (2022). https://doi.org/10.1007/s41095-021-0237-5

Download citation

  • Received: 20 March 2021

  • Accepted: 29 April 2021

  • Published: 06 December 2021

  • Issue Date: June 2022

  • DOI: https://doi.org/10.1007/s41095-021-0237-5

Share this article

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

Keywords

  • nose shape recovery
  • single image 3D reconstruction
  • contour correspondence
  • Laplacian deformation
Download PDF

Working on a manuscript?

Avoid the common mistakes

Advertisement

Over 10 million scientific documents at your fingertips

Switch Edition
  • Academic Edition
  • Corporate Edition
  • Home
  • Impressum
  • Legal information
  • Privacy statement
  • California Privacy Statement
  • How we use cookies
  • Manage cookies/Do not sell my data
  • Accessibility
  • FAQ
  • Contact us
  • Affiliate program

Not affiliated

Springer Nature

© 2023 Springer Nature Switzerland AG. Part of Springer Nature.