Soppari, Kavitha and Kavya, Pakide and Teja, Kotla Pranay and Sai, Bethi Pavan (2025) A survey on image captioning methods. World Journal of Advanced Research and Reviews, 26 (2). pp. 3134-3143. ISSN 2581-9615
![WJARR-2025-1705.pdf [thumbnail of WJARR-2025-1705.pdf]](https://eprint.scholarsrepository.com/style/images/fileicons/text.png)
WJARR-2025-1705.pdf - Published Version
Available under License Creative Commons Attribution Non-commercial Share Alike.
Abstract
Image captioning is a task that Involves Natural Language Processing concepts to recognize the context of an image and describe them in a natural language like English. It requires good knowledge of Deep learning. Python, working on Jupyter notebooks, Keras library, Numpy, and Natural language processing It is a Python based project where we will use deep learning techniques of Convolutional Neural Networks and a type of Recurrent Neural Network (LSTM) together. The biggest challenge is most definitely being able to create a description that must capture not only the objects contained in an image, but also express how these objects relate to each other. Automatically describing the content of an image is a fundamental problem in artificial intelligence that connects computer vision and natural language processing here, we present a generative model based on a deep recurrent architecture that combines recent advances in computer vision and machine translation and that can be used to generate natural sentences describing an image. It could have great impact, for instance by helping visually impaired people better understand the content of images on the web.
Item Type: | Article |
---|---|
Official URL: | https://doi.org/10.30574/wjarr.2025.26.2.1705 |
Uncontrolled Keywords: | CNN; LSTM; Image detection; Deep learning; Natural Language Processing |
Depositing User: | Editor WJARR |
Date Deposited: | 20 Aug 2025 11:35 |
Related URLs: | |
URI: | https://eprint.scholarsrepository.com/id/eprint/3363 |