Skip to main content

Deep Learning Techniques for Automated Image Captioning

  • Conference paper
  • First Online:
Smart Trends in Computing and Communications

Part of the book series: Lecture Notes in Networks and Systems ((LNNS,volume 286))

  • 802 Accesses

Abstract

Automated Image Captioning involves understanding the semantic information of an image and expressing it in natural language. Among the many approaches proposed, deep learning-based techniques have achieved state-of-the-art results in solving this problem. In this paper, three primary, distinct deep learning-based approaches to solve this problem are introduced and compared: encoder-decoder frameworks, neuroevolution, and attention-based approaches. This paper covers their mechanisms and their performance, and highlights where they differ from each other. To conclude, the results of these approaches on benchmark dataset and metrics are presented.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 189.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 249.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Fei-Fei L, Iyer A, Koch C, Perona P (2007) What do we perceive in a glance of a real-world scene? J Vis 7(1):10

    Article  Google Scholar 

  2. Sharma H, Agrahari M, Singh SK, Firoj M, Mishra RK (2020) Image captioning: a comprehensive survey. In: 2020 international conference on power electronics & IoT applications in renewable energy and its control (PARC). IEEE, pp 325–328

    Google Scholar 

  3. Bai S, An S (2018) A survey on automatic image caption generation. Neurocomputing 311:291–304

    Article  Google Scholar 

  4. Wang X, Zhao Y, Pourpanah F (2020) Recent advances in deep learning

    Google Scholar 

  5. Kiros R, Salakhutdinov R, Zemel, RS (2014) Unifying visual-semantic embeddings with multimodal neural language models

    Google Scholar 

  6. Vinyals O, Toshev A, Bengio S, Erhan D, Show and tell: a neural image caption generator. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 3156–3164

    Google Scholar 

  7. Wu Q, Shen C, Liu L, Dick A, Van Den Hengel A, What value do explicit high level concepts have in vision to language problems? In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 203–212

    Google Scholar 

  8. Wei Y, Xia W, Huang J, Ni B, Dong J, Zhao Y, Yan S, CNN: single-label to multi-label

    Google Scholar 

  9. Stanley KO, Miikkulainen R, Evolving neural networks through augmenting topologies, 10(2):99–127

    Google Scholar 

  10. Miikkulainen R, Liang J, Meyerson E, Rawal A, Fink D, Francon O, Raju B, Shahrzad H, Navruzyan A, Duffy N, Hodjat B, Evolving deep neural networks

    Google Scholar 

  11. Spratling MW, Johnson MH (2004) A feedback model of visual attention. J Cogn Neurosci 16(2):219–237

    Article  Google Scholar 

  12. Mnih V, Heess N, Graves A, Kavukcuoglu K (2014) Recurrent models of visual attention. arXiv preprint arXiv:1406.6247

  13. Ba J, Mnih V, Kavukcuoglu K (2014) Multiple object recognition with visual attention, arXiv preprint arXiv:1412.7755

  14. Xu K, Ba J, Kiros R, Cho K, Courville A, Salakhutdinov R, Zemel R, Bengio Y (2016) Show, attend and tell: neural image caption generation with visual attention

    Google Scholar 

  15. Yu J, Li J, Yu Z, Huang Q (2019) Multimodal transformer with multi-view visual representation for image captioning. IEEE Trans Circ Syst Video Technol 30(12):4467–4480

    Article  Google Scholar 

  16. Wolf T, Chaumond J, Debut L, Sanh V, Delangue C, Moi A, Cistac P, Funtowicz M, Davison J, Shleifer S et al (2020) Transformers: state-of-the-art natural language processing. In: Proceedings of the 2020 conference on empirical methods in natural language processing: system demonstrations, pp 38–45

    Google Scholar 

  17. Carion N, Massa F, Synnaeve G, Usunier N, Kirillov A, Zagoruyko S (2020) End-to-end object detection with transformers. In: European conference on computer vision. Springer, pp 213–229

    Google Scholar 

  18. Papineni K, Roukos S, Ward T, Zhu W-J (2002) Bleu: a method for automatic evaluation of machine translation. In: Proceedings of the 40th annual meeting of the association for computational linguistics, pp 311–318

    Google Scholar 

  19. Banerjee S, Lavie A (2005) Meteor: an automatic metric for MT evaluation with improved correlation with human judgments. In: Proceedings of the ACL workshop on intrinsic and extrinsic evaluation measures for machine translation and/or summarization, pp 65–72

    Google Scholar 

  20. Vedantam R, Zitnick CL, Parikh D (2015) Cider: consensus-based image description evaluation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4566–4575

    Google Scholar 

  21. Chen X, Fang H, Lin T-Y, Vedantam R, Gupta S, Dollar P, Zitnick CL, Microsoft COCO captions: data collection and evaluation server, Version: 2

    Google Scholar 

  22. Donahue J, Hendricks LA, Guadarrama S, Rohrbach M, Venugopalan S, Saenko K, Darrell T, Long-term recurrent convolutional networks for visual recognition and description. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2625–2634

    Google Scholar 

  23. Yang Z, Yuan Y, Wu Y, Salakhutdinov R, Cohen WW (2016) Review networks for caption generation. arXiv preprint arXiv:1605.07912

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2022 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Srivastava, S., Chaudhari, Y., Damania, Y., Jadhav, P. (2022). Deep Learning Techniques for Automated Image Captioning. In: Zhang, YD., Senjyu, T., So-In, C., Joshi, A. (eds) Smart Trends in Computing and Communications. Lecture Notes in Networks and Systems, vol 286. Springer, Singapore. https://doi.org/10.1007/978-981-16-4016-2_55

Download citation

Publish with us

Policies and ethics