Abstract
Reverse image search is a content-based image retrieval (CBIR) query technique that takes a sample image as an input, and search is performed based on it. Reverse image search is characterized by a lack of search terms. Using reverse image search, one can find the original source of images, find plagiarized photos, detect fake accounts on social media, etc. Reverse image search works by uploading an image by the user, and searching of images is carried out by using the corresponding meta tags, HTML tags or color distributions of the image. The search engine currently functions by using this information and not the context of the image, thus resulting in inaccurate outcomes. Labeling the image using keywords relevant to the image is a very generic explanation to an image and can lead to ambiguous interpretations. This can be overcome by giving a more specific description to an image, thereby preventing misinterpretations and providing more specific results. The proposed system aims at optimizing the search results by depicting more clearly the relationships between the objects in an image with the help of image captioning.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Elamri C, de Planque T (2016) Automated neural image caption generator for visually impaired people. Department of Computer Science, Stanford University. https://cs224d.stanford.edu/reports/mcelamri.pdf. Accessed 18 Oct 2018
He X, Shi B, Bai X, Xia G-S, Zhang Z, Dong W (2016) Image caption generation with part of speech guidance. In: IEEE conference on computer vision and pattern recognition 2016, school of electronic information and communications, Huazhong University
Yang Z, Zhang Y-J, ur Rehman S, Huang Y (2016) Image caption with object detection and localization. Department of Electronic Engineering, Tsinghua University, Beijing 100084, China. https://arxiv.org/pdf/1706.02430.pdf
Zhou L, Xu C, Koch P, Corse JJ (2017) Image caption generation with text-conditional semantic attention. Robotics Institute, University of Michigan, Department of Computer Science, University of Rochester, Electrical and Computer Engineering, University of Michigan, 2017
Asakawa S, Ogata T (2017) Comparison between variational autoencoder and encoder-decoder models for short conversation. Proc Int Conf Artif Life Robot 22:639–642. https://doi.org/10.5954/ICAROB.2017.OS1-4
Chen J, Dong W, Li M (2017) Image caption generator based on deep neural networks. Department of Computer Science, University of British Columbia. https://www.cs.ubc.ca/~carenini/TEACHING/CPSC503-19/FINAL-PROJECTS-2016/image_caption_generator_final_report.pdf. Accessed 12 Nov 2018
Karpathy A, Fei-Fei L (2015) Deep visual-semantic alignments for generating image descriptions. In: IEEE conference on computer vision and pattern recognition
Khurana K, Awasthi R (2013) Techniques for object recognition in images and multi-object detection. Int J Adv Res Comp Eng Technol (IJARCET) 2(4):1383
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Kansara, D., Shinde, A., Suba, Y., Joshi, A. (2020). Optimizing Reverse Image Search by Generating and Assigning Suitable Captions to Images. In: Vasudevan, H., Michalas, A., Shekokar, N., Narvekar, M. (eds) Advanced Computing Technologies and Applications. Algorithms for Intelligent Systems. Springer, Singapore. https://doi.org/10.1007/978-981-15-3242-9_59
Download citation
DOI: https://doi.org/10.1007/978-981-15-3242-9_59
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-15-3241-2
Online ISBN: 978-981-15-3242-9
eBook Packages: EngineeringEngineering (R0)