Sign In to Follow Application
View All Documents & Correspondence

Recommendations For Generating Automatic Captions Based On Visual Content

Abstract: RECOMMENDATIONS FOR GENERATING AUTOMATIC CAPTIONS BASED ON VISUAL CONTENT An automatic image caption generation system that integrates computer vision and natural language processing techniques to produce accurate and contextually meaningful textual descriptions for images. The system utilizes Convolutional Neural Networks (CNNs) for feature extraction and Long Short-Term Memory (LSTM) networks for caption generation. Word embeddings such as Word2Vec and GloVe enhance semantic understanding, while large-scale datasets like Microsoft COCO improve accuracy. The model incorporates attention mechanisms to prioritize significant image features and relationship detection to contextualize objects within a scene. Reinforcement learning techniques optimize caption quality over time. This technology benefits visually impaired users, enhances search engine indexing, and improves content tagging in digital platforms. By effectively capturing and translating visual information into coherent language, the system outperforms existing approaches in generating fluid, human-like image descriptions.

Get Free WhatsApp Updates!
Notices, Deadlines & Correspondence

Patent Information

Application #
Filing Date
19 February 2025
Publication Number
10/2025
Publication Type
INA
Invention Field
COMPUTER SCIENCE
Status
Email
Parent Application

Applicants

SR UNIVERSITY
ANANTHSAGAR, HASANPARTHY (M), WARANGAL URBAN, TELANGANA - 506371, INDIA

Inventors

1. MR. P. RADHAKRISHNAN
SR UNIVERSITY, ANANTHASAGAR, HASANPARTHY (M), WARANGAL URBAN, TELANGANA - 506371, INDIA
2. DR. N. SHARMILA BANU
SR UNIVERSITY, ANANTHASAGAR, HASANPARTHY (M), WARANGAL URBAN, TELANGANA - 506371, INDIA
3. DR. P. PRAVEEN
SR UNIVERSITY, ANANTHASAGAR, HASANPARTHY (M), WARANGAL URBAN, TELANGANA - 506371, INDIA
4. DR. VISWANATH BIJALWAN
SR UNIVERSITY, ANANTHASAGAR, HASANPARTHY (M), WARANGAL URBAN, TELANGANA - 506371, INDIA
5. MR. DUGYALA VINAY
SR UNIVERSITY, ANANTHASAGAR, HASANPARTHY (M), WARANGAL URBAN, TELANGANA - 506371, INDIA

Specification

Description:FIELD OF THE INVENTION
This invention relates to Recommendations for Generating Automatic Captions based on Visual Content
BACKGROUND OF THE INVENTION
The problem statement of automatic image caption generation is to produce accurate and meaningful descriptions for images through the application of NLP and Computer Vision. The challenge is to examine the visual content of the image and transform it into accurate and meaningful captions. The aim is to ensure that the captions are not only accurate but also fluent and contextually aware.
The existing systems for automatic caption generation depends on basic image processing that neglect a few important details and relationships with in images These existing systems generate simple captions that does not consider in capturing the visual content. The proposed system uses advanced machine learning techniques such as CNN for efficient feature extraction and LSTM networks to generate contextual sentences. This combination improves image understanding which results in more accurate and detailed captions that reflect object relationships.
Automatic caption generation represents a significant development in computer vision and NLP. The existing methods frequently found to be difficult to complex scene. However by integrating CNN and LSTM the proposed system effectively addresses these challenges and more effectively translating visual scene into language. Using a large training datasets and advanced word embedding techniques, it helps to produce coherent and relevant captions. It improves the quality of the description as well as it expands the potential applications including assistance for the visually impaired or improving search engines. In conclusion this method represents significant progress in connecting visual understanding with language which will ultimately lead to better AI solutions.
SUMMARY OF THE INVENTION
This summary is provided to introduce a selection of concepts, in a simplified format, that are further described in the detailed description of the invention.
This summary is neither intended to identify key or essential inventive concepts of the invention and nor is it intended for determining the scope of the invention.
To further clarify advantages and features of the present invention, a more particular description of the invention will be rendered by reference to specific embodiments thereof, which is illustrated in the appended drawings. It is appreciated that these drawings depict only typical embodiments of the invention and are therefore not to be considered limiting of its scope. The invention will be described and explained with additional specificity and detail with the accompanying drawings.
Automatic image caption generation is a complicated process that combines computer vision and NLP to produce descriptive text for the given image. The complexity in extracting the features, classifying the objects and generating contextually relevant to the images are risky. The system must effectively translate visual information into coherent sentences.
Automatic image captioning is a multidisciplinary domain that utilizing both computer vision and NLP techniques. Convolution Neural Networks are utilized to analyze the images and identify critical visual features. An LSTM model predicts the words from visual context to generate the captions. The objective is to produce captions that are also fluid and meaningful. This technology is beneficial for visually impaired people and improves search engines and content tagging system.
BRIEF DESCRIPTION OF THE DRAWINGS
The illustrated embodiments of the subject matter will be understood by reference to the drawings, wherein like parts are designated by like numerals throughout. The following description is intended only by way of example, and simply illustrates certain selected embodiments of devices, systems, and methods that are consistent with the subject matter as claimed herein, wherein:
FIGURE 1: SYSTEM ARCHITECTURE
The figures depict embodiments of the present subject matter for the purposes of illustration only. A person skilled in the art will easily recognize from the following description that alternative embodiments of the structures and methods illustrated herein may be employed without departing from the principles of the disclosure described herein.
DETAILED DESCRIPTION OF THE INVENTION
The detailed description of various exemplary embodiments of the disclosure is described herein with reference to the accompanying drawings. It should be noted that the embodiments are described herein in such details as to clearly communicate the disclosure. However, the amount of details provided herein is not intended to limit the anticipated variations of embodiments; on the contrary, the intention is to cover all modifications, equivalents, and alternatives falling within the scope of the present disclosure as defined by the appended claims.
It is also to be understood that various arrangements may be devised that, although not explicitly described or shown herein, embody the principles of the present disclosure. Moreover, all statements herein reciting principles, aspects, and embodiments of the present disclosure, as well as specific examples, are intended to encompass equivalents thereof.
The terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of example embodiments. As used herein, the singular forms “a",” “an” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will be further understood that the terms “comprises,” “comprising,” “includes” and/or “including,” when used herein, specify the presence of stated features, integers, steps, operations, elements and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components and/or groups thereof.
It should also be noted that in some alternative implementations, the functions/acts noted may occur out of the order noted in the figures. For example, two figures shown in succession may, in fact, be executed concurrently or may sometimes be executed in the reverse order, depending upon the functionality/acts involved.
In addition, the descriptions of "first", "second", “third”, and the like in the present invention are used for the purpose of description only, and are not to be construed as indicating or implying their relative importance or implicitly indicating the number of technical features indicated. Thus, features defining "first" and "second" may include at least one of the features, either explicitly or implicitly.
Unless otherwise defined, all terms (including technical and scientific terms) used herein have the same meaning as commonly understood by one of ordinary skill in the art to which example embodiments belong. It will be further understood that terms, e.g., those defined in commonly used dictionaries, should be interpreted as having a meaning that is consistent with their meaning in the context of the relevant art and will not be interpreted in an idealized or overly formal sense unless expressly so defined herein.
Automatic image caption generation is a complicated process that combines computer vision and NLP to produce descriptive text for the given image. The complexity in extracting the features, classifying the objects and generating contextually relevant to the images are risky. The system must effectively translate visual information into coherent sentences.
Automatic image captioning is a multidisciplinary domain that utilizing both computer vision and NLP techniques. Convolution Neural Networks are utilized to analyze the images and identify critical visual features. An LSTM model predicts the words from visual context to generate the captions. The objective is to produce captions that are also fluid and meaningful. This technology is beneficial for visually impaired people and improves search engines and content tagging system.
This approach combines advanced computer vision and NLP processing methods to produce accurate and complete image caption. This system can identify the objects and also capable of organizing their relationships in a visual by employing Convolution Neural Networks for feature extraction and LSTM networks for sequence generation. This produces more contextual and meaningful captions than existing approach.
Using word embeddings like Word2Vec and Glove as well as large datasets like Microsoft COCO further improves the system’s capacity to produce natural and human like descriptions. This model is generating captions for complex visual by understanding both the relationships between the objects and collaborative functions. This makes an effective tool for applications such as content tagging and improved search capabilities.

, Claims:1. A system for automatic image caption generation, comprising: a. an image processing module utilizing a Convolutional Neural Network (CNN) to extract visual features from an input image; b. a natural language processing module employing a Long Short-Term Memory (LSTM) network to generate textual descriptions based on the extracted features; c. a word embedding module configured to enhance semantic understanding using pre-trained word embeddings such as Word2Vec or GloVe; d. a dataset training module using large datasets, including Microsoft COCO, to improve model accuracy; e. a caption generation module that synthesizes the extracted features into a coherent and contextually relevant textual description.
2. The system as claimed in claim 1, wherein the CNN is pre-trained on an image classification dataset to improve feature extraction accuracy.
3. The system as claimed in claim 1, wherein the LSTM model is trained with attention mechanisms to enhance the contextual accuracy of generated captions.
4. The system as claimed in claim 1, further comprising a relationship detection module configured to identify spatial and functional relationships between detected objects within an image.
5. The system as claimed in claim 1, wherein the dataset training module continuously updates the model using reinforcement learning techniques to improve caption accuracy over time.
6. A method for generating automatic image captions, comprising: a. receiving an input image; b. extracting visual features using a Convolutional Neural Network; c. processing the extracted features through an LSTM-based caption generation model; d. enhancing semantic relevance using word embeddings; e. generating a structured and contextually relevant caption based on identified objects and their relationships.
7. The method as claimed in claim 6, wherein the system employs an attention mechanism within the LSTM model to focus on significant objects within the image.
8. The method as claimed in claim 6, wherein the system uses pre-trained datasets such as Microsoft COCO to improve caption generation accuracy.
9. The method as claimed in claim 6, further comprising fine-tuning the captioning model using reinforcement learning based on user feedback.
10. The method as claimed in claim 6, wherein the generated caption is further refined using a grammar correction model to enhance fluency and readability.

Documents

Application Documents

# Name Date
1 202541014308-STATEMENT OF UNDERTAKING (FORM 3) [19-02-2025(online)].pdf 2025-02-19
2 202541014308-REQUEST FOR EARLY PUBLICATION(FORM-9) [19-02-2025(online)].pdf 2025-02-19
3 202541014308-POWER OF AUTHORITY [19-02-2025(online)].pdf 2025-02-19
4 202541014308-FORM-9 [19-02-2025(online)].pdf 2025-02-19
5 202541014308-FORM FOR SMALL ENTITY(FORM-28) [19-02-2025(online)].pdf 2025-02-19
6 202541014308-FORM 1 [19-02-2025(online)].pdf 2025-02-19
7 202541014308-EVIDENCE FOR REGISTRATION UNDER SSI(FORM-28) [19-02-2025(online)].pdf 2025-02-19
8 202541014308-EVIDENCE FOR REGISTRATION UNDER SSI [19-02-2025(online)].pdf 2025-02-19
9 202541014308-EDUCATIONAL INSTITUTION(S) [19-02-2025(online)].pdf 2025-02-19
10 202541014308-DRAWINGS [19-02-2025(online)].pdf 2025-02-19
11 202541014308-DECLARATION OF INVENTORSHIP (FORM 5) [19-02-2025(online)].pdf 2025-02-19
12 202541014308-COMPLETE SPECIFICATION [19-02-2025(online)].pdf 2025-02-19