Enhanced Image Captioning Using CNN and Transformers with Attention Mechanism

International Journal of Engineering Innovations and Management Strategies 1 (1):1-12 (2024)
  Copy   BIBTEX

Abstract

Image captioning has seen remarkable advancements with the integration of deep learning techniques, notably Convolutional Neural Networks (CNNs) and Long Short-Term Memory (LSTM) networks, for generating descriptive captions for images. Despite these improvements, capturing intricate details and context remains a challenge. This project introduces an enhanced image captioning model that integrates transformers with an attention mechanism to address these limitations. By leveraging CNNs for feature extraction and LSTMs for sequence generation, while utilizing transformers to apply sophisticated attention to significant image regions, the proposed model aims to generate more contextually rich and coherent captions. Experimental results indicate that incorporating transformers with attention mechanisms leads to a significant enhancement in caption accuracy and descriptiveness, surpassing traditional CNN-LSTM models. This advancement is particularly beneficial in various applications, including assistive technologies for the visually impaired, content-based image retrieval systems, automatic image annotation for digital asset management, and improved human-computer interaction. This approach represents a substantial step forward in achieving more precise and detailed image captioning, with potential impacts across numerous fields.

Analytics

Added to PP
today

Downloads
0

6 months
0

Historical graph of downloads since first upload

Sorry, there are not enough data points to plot this chart.
How can I increase my downloads?