Sign In to Follow Application
View All Documents & Correspondence

Multimodal Language Translation And Communication Wearable Glasses

Abstract: A multimodal language translation and communication wearable glasses, comprising a transparent frame 101 developed to be worn by a user and adapted to rest on user's nose and ears, an array of directional microphones 106 configured to capture ambient sound from multiple directions, an AI-powered camera 110 integrated with facial recognition and gesture recognition capabilities, an augmented reality heads-up display (HUD) integrated into lens 103 of the frame 101, a translation module configured to provide real-time translation of spoken language, sign language, and gestures into a preferred language of user, at least two ear units 107 mounted on articulated arms connected to temple ends to enhance audio clarity and filter background noise, and a microcontroller configured to coordinate data processing from microphones 106, AI camera, sensors, translation module, and AR HUD, microcontroller facilitates real-time context-aware language translation and communication assistance for user interaction.

Get Free WhatsApp Updates!
Notices, Deadlines & Correspondence

Patent Information

Application #
Filing Date
10 June 2025
Publication Number
27/2025
Publication Type
INA
Invention Field
PHYSICS
Status
Email
Parent Application

Applicants

Marwadi University
Rajkot - Morbi Road, Rajkot 360003 Gujarat, India.

Inventors

1. Fenil Vadher
Department of Information and Communication Technology, Marwadi University, Rajkot - Morbi Road, Rajkot 360003 Gujarat, India.
2. Chandrasinh D Parmar
Department of Information and Communication Technology, Marwadi University, Rajkot - Morbi Road, Rajkot 360003 Gujarat, India.
3. Nishith Kotak
Department of Information and Communication Technology, Marwadi University, Rajkot - Morbi Road, Rajkot 360003 Gujarat, India.

Specification

Description:FIELD OF THE INVENTION

[0001] The present invention relates to a wearable device in the form of eyewear. More specifically, the present invention is directed towards a multimodal language translation and communication wearable glasses that is capable of helping users to communicate with others in their preferred languages or through non-verbal signals to support clear communication in diverse situations.

BACKGROUND OF THE INVENTION

[0002] Language barriers and communication difficulties are common in today’s global environment, where people from different regions, cultures, and abilities regularly interact. Whether in travel, education, healthcare, or daily social encounters, there is a growing need for tools that support clear and effective communication between individuals who speak different languages or use alternative communication methods such as sign language or gestures. As society becomes more connected and inclusive, the demand for accessible and efficient translation and communication tools continues to rise.

[0003] Traditionally, communication between individuals speaking different languages has relied on handheld translation devices, mobile apps, or written guides. These solutions often require users to manually input or speak words into a device, wait for translation, and then relay the response. Sign language interpretation typically requires a human interpreter or limited software-based apps that rely on a mobile phone camera. These methods interrupt the natural flow of conversation, lack real-time accuracy, and are inconvenient in dynamic or fast-paced settings. In addition, they often do not support multimodal inputs such as gestures or facial expressions. Furthermore, most available solutions fail to offer personalized experiences based on a user’s preferences or location context.

[0004] US20200074745A1 discloses a method detects a real object with a wearable electronic device and displays a virtual image of the real object with the real object on a display of the wearable electronic device. The wearable electronic device displays movement of the virtual image of the real object to show a task to be completed and detects a completion of the task.

[0005] WO2020238476A1 discloses the augmented reality glasses comprise: a main lens; a virtual imaging module, which is used for imaging on the main lens; and a zoom correction module, which is stacked on the main lens and located in an imaging region of the virtual imaging module, wherein the zoom correction module comprises a liquid zoom lens and a pair of light-transmitting electrodes, and the pair of light-transmitting electrodes is provided on opposite surfaces of the liquid zoom lens and is used for generating a variable voltage that causes the liquid zoom lens to deform.

[0006] Conventionally, many devices have been developed in order to assist users in overcoming language differences and sensory limitations. However, the devices mentioned in the prior arts have limitations pertaining to user adaptability, multimodal input recognition, and responding by detecting real-time situations. Most rely on handheld input, lack continuous and hands-free operation, and do not integrate spatial awareness, fail to support gesture or facial recognition with precision.

[0007] In order to overcome the aforementioned drawbacks, there exists a need in the art to develop a wearable glass that requires to be capable of interpreting verbal and non-verbal communication cues in view of providing real-time translation and environmental awareness to the user in order to facilitate intuitive and inclusive interaction in diverse and multilingual settings.

OBJECTS OF THE INVENTION

[0008] The principal object of the present invention is to overcome the disadvantages of the prior art.

[0009] An object of the present invention is to develop a wearable glass that allows users to interact effectively with people speaking different languages by translating speech, signs, and gestures into a preferred language instantly, improving communication in multicultural or multilingual environments.

[0010] Another object of the present invention is to develop a wearable glass that is capable of assisting users in communicating through multiple modes for better accessibility for individuals with hearing or speech impairments, thus enhancing inclusivity in everyday communication.

[0011] Another object of the present invention is to develop a wearable glass that filters out background noise to enhance primary speaker’s voice, thus allowing users to receive clear audio and accurate translations in environments with heavy sound interference or distractions.

[0012] Another object of the present invention is to develop a wearable glass that is capable of detecting relevant visual cues and surroundings for helping the user to stay aware of any hazards nearby in unfamiliar or dynamic environments.

[0013] Yet another object of the present invention is to develop a wearable glass evice that is capable of operating intuitively without manual input, adapting to user behavior, preferences, and contextual settings for enhancing user convenience.

[0014] The foregoing and other objects, features, and advantages of the present invention will become readily apparent upon further review of the following detailed description of the preferred embodiment as illustrated in the accompanying drawings.

SUMMARY OF THE INVENTION

[0015] The present invention relates to a multimodal language translation and communication wearable glasses that is capable of assisting users to understand and respond to any spoken language, gestures, and surroundings by translating the language and cues, for enhanced communication and awareness in public, personal, or professional settings.

[0016] According to an embodiment of the present invention, a multimodal language translation and communication wearable glasses is disclosed comprising a transparent frame developed to be worn by a user and adapted to rest on the user’s nose and ears, an array of directional microphones configured to capture ambient sound from multiple directions, an AI (artificial intelligence)-powered camera integrated with facial recognition and gesture recognition capabilities, an augmented reality heads-up display (HUD) integrated into a pair electrochromic lens provided with the frame, a translation module configured to provide real-time translation of spoken language, sign language, and gestures into a preferred language of the user, atleast two ear units mounted on articulated arms connected to the temple ends to enhance audio clarity and filter background noise, a microcontroller configured to coordinate data processing from the microphones, AI camera, sensors, translation module, and AR HUD, an outgoing microphone is configured to capture the user’s voice pitch and provide haptic feedback alerts if the pitch is outside of a predefined range.

[0017] According to another embodiment of the present invention, the present invention further includes a liquid crystal lens is housed in a concealed motorized vertical slider above the frame configured to move downward in front of the user’s eyes, a curved-shaped cover flap is attached to the outer rims of the lenses via motorized hinge joints, configured to automatically close and protect the lenses, nanostructured capacitive sensors are provided around the lens edges to track eye movements of both the user and speaker, a GPS (Global Positioning System) module is integrated with the microcontroller tracks real-time reporting location, a dust sensor integrated to monitor lens surfaces and triggers a plurality of micro air-jet nozzles positioned along the inner and outer upper frame to blow away dust or debris, and a user-interface is inbuilt in a computing unit, allowing user(s) to manage event schedules, maps, translation preferences, and hearing or vision profiles securely stored in a connected database.

[0018] While the invention has been described and shown with particular reference to the preferred embodiment, it will be apparent that variations might be possible that would fall within the scope of the present invention.

BRIEF DESCRIPTION OF THE DRAWINGS

[0019] These and other features, aspects, and advantages of the present invention will become better understood with regard to the following description, appended claims, and accompanying drawings where:
Figure 1 illustrates an isometric view of a multimodal language translation and communication wearable glasses; and
Figure 2 illustrates a perspective view of the wearable glasses having a flap in a deployed state.

DETAILED DESCRIPTION OF THE INVENTION

[0020] The following description includes the preferred best mode of one embodiment of the present invention. It will be clear from this description of the invention that the invention is not limited to these illustrated embodiments but that the invention also includes a variety of modifications and embodiments thereto. Therefore, the present description should be seen as illustrative and not limiting. While the invention is susceptible to various modifications and alternative constructions, it should be understood, that there is no intention to limit the invention to the specific form disclosed, but, on the contrary, the invention is to cover all modifications, alternative constructions, and equivalents falling within the spirit and scope of the invention as defined in the claims.

[0021] In any embodiment described herein, the open-ended terms "comprising," "comprises,” and the like (which are synonymous with "including," "having” and "characterized by") may be replaced by the respective partially closed phrases "consisting essentially of," consists essentially of," and the like or the respective closed phrases "consisting of," "consists of, the like.

[0022] As used herein, the singular forms “a,” “an,” and “the” designate both the singular and the plural, unless expressly stated to designate the singular only.

[0023] The present invention relates to a multimodal language translation and communication wearable glasses that is developed to improve communication for users by processing speech, signs, visual cues or audio inputs, thus providing support in real-time understanding and response especially in diverse areas speaking different languages, or in noisy, or unfamiliar environments.

[0024] Referring to Figure 1 and 2, an isometric view of a multimodal language translation and communication wearable glasses and a perspective view of the wearable glasses having a flap in a deployed state are illustrated, respectively, comprises of a transparent frame 101 developed to be worn by a user constructed with a pair of lens holding rims 102, including an electrochromic lens 103, motorized nose bridge 104, telescopic temple arms 105, an array of directional microphones 106 configured with frame 101, two ear units 107 mounted on each articulated arms connected to the temple arms 105, a liquid crystal lens 108 is housed in a concealed motorized vertical slider 109 above the frame 101.

[0025] Figure1 and 2 further illustrates an AI (artificial intelligence)-powered camera 110 configured on the frame 101, an outgoing microphone 111 installed with the frame 101, nose bridge 104 is attached with a pair of nose pads 112, and a curved-shaped cover flap 201 is attached to the outer rims 102 of the lenses 103 via motorized hinge joints 202.

[0026] The present invention includes a transparent frame 101 preferably in a shape of a spectacles that is developed to to rest on the user’s nose and ears. The frame 101 is linked with a microcontroller for processing data from various components of the glasses. The microcontroller, mentioned herein, is preferably an Arduino microcontroller. The Arduino microcontroller used herein controls the overall functionality of the components linked to it. The Arduino microcontroller is an open-source programming platform. The microcontroller receives the data from various electronic units and generates a command signal for further processing.

[0027] In an embodiment of the present invention, the microcontroller is linked with a push button installed on the glasses. The push button is accessed by the user, and as the button is pressed, it opens up an electrical circuit that allows currents to flow for powering an associated microcontroller of the glasses for operating of all the linked components for performing their respective functions upon actuation.

[0028] The frame 101 includes a pair of lens-holding rims 102 that are in the shape of circular or oval-shaped sections. The rims 102 provide a stable and secure hold to a pair of lenses 103 installed with the rims. These rims 102 are typically positioned in front of each eye and are integrated into the main structure of the glasses frame 101.

[0029] The lens 103 herein is electrochromic lens 103 that automatically adjust tint based on ambient light intensity. Electrochromic lenses 103 work by using electrically responsive materials that change their tint or transparency when a small voltage is applied. These lenses 103 consist of multiple layers, including an electrochromic layer that is the active material that changes color or tint, a conductive layer to apply electrical current and an ion-conducting layer that allows ions to move between layers when voltage is applied.

[0030] The lighting conditions are detected by light sensors integrated with the lens. The sensor is typically a photodiode that detects light by change in the resistance based on the brightness of the light. This signal is sent to the microcontroller, which interprets the level of light. If the light intensity is high, the microcontroller triggers the electrochromic lens 103 to darken by applying a low electrical voltage. If the light is low, the signal prompts the lens 103 to return to a clearer state.

[0031] When ambient light intensity increases (e.g., bright sunlight), the sensor sends a signal to the microcontroller. The microcontroller sends the commands to conductive layer to apply a low voltage to the lens. This causes ions to move into the electrochromic layer, darkening the lens 103 to reduce glare and improve visual comfort. When light intensity decreases, the process is reversed, and the lens 103 becomes clear again. This reaction is gradual, reversible, and helps protect the user's eyes while maintaining visibility.

[0032] The frame 101 is integrated with plurality of optical sensors for detecting facial features of the user to adjust the frame 101 according to the facial pattern. The optical sensors work by emitting light usually infrared and detecting the reflection from the user’s face to capture detailed information about facial features. These sensors scan key points such as the nose bridge 104, cheekbones, and ear positions to create a facial map. The reflected light is converted into electrical signals, which are processed by a microcontroller or processor to analyse the shape, size, and contours of the face.

[0033] The rims 102 have a gap in between that forms a nose bridge 104 that is motorized to adjust automatically customized to the user’s nasal dimension, ensuring comfort during long wearing hours. The motorized nose bridge 104 works by using small electric motors connected to the two rim sections that form the bridge 104. When the user wears the glasses, sensors detect the nasal dimensions and send this information to the microcontroller. The microcontroller then controls the motors to move the rims 102 closer together or farther apart, adjusting the width and shape of the nose bridge 104 automatically. This precise adjustment ensures that the glasses fit snugly and comfortably on the user’s nose, reducing pressure points and enhancing comfort during extended use. The nose bridge 104 is attached with a pair of nose pads 112 that are equipped with cushioning material to provide better comfort and prevent strain on the nasal area.

[0034] The frame 101 is installed with a telescopic temple arms 105 that are connected to each end of the frame 101. The arm 105 extend/retracts to adjust the length to ensure a comfortable and adjustable fit for diverse users. The microcontroller sends a signal to a pneumatic unit associated with the temple arms 105 that includes an air compressor, air cylinder, air valves and piston which works in collaboration to aid in extension and retraction of the temple arms 105.

[0035] The pneumatic unit is operated by the microcontroller, such that the microcontroller actuates valve to allow passage of compressed air from the compressor within the cylinder from one end, the compressed air further develops pressure against the piston and results in pushing and extending the piston. The piston is connected with the temple arms 105 and due to applied pressure the temple arms 105 extends and similarly, the microcontroller retracts the arms 105 by pushing compressed air via the other end of the cylinder, by opening the corresponding valve resulting in retraction of the piston, and the retraction of the temple arms 105 to configure to the user’s face pattern for providing comfort.

[0036] The telescopic segments of the temple arms 105 are connected together via ball and socket joints for providing flexible movement and better adjustment to the user. The microcontroller sends a signal to a motor connected to the ball and socket joint that provides a 360-degree rotation to the temple arms 105 at a desired angle for flexible movement and better adjustment. The ball and socket joint is a coupling consisting of a ball joint securely locked within a socket joint, where the ball joint is able to move in a 360-dgree rotation within the socket thus, providing the required rotational motion to the temple arms 105. The ball and socket joint is powered by a DC (direct current) motor that is actuated by the microcontroller thus providing multidirectional movement to the temple arms 105.

[0037] Piezo-resistive sensors are embedded within the temple structure to detect any strain or pressure. Piezo-resistive sensors work by changing their electrical resistance when mechanical strain, such as pressure or deformation, is applied to them. The sensors experience changes in length and cross-sectional area under strain, which alters their resistance. This variation is detected electrically, often through a circuit and then the mechanical strain into a measurable electrical signal, allowing the sensor to detect and quantify pressure or stress in real time.

[0038] The inner side of the temple arms 105 (facing the user) is equipped with motorized foam rollers located at the end sections. These rollers are activated in case of stress to adjust their position for user comfort. The motorized foam rollers work by using small electric motors to adjust their position automatically when stress or pressure is detected, allowing them to gently apply or relieve pressure for improved user comfort. When sensors detect strain or discomfort, the motors drive the foam rollers to move or rotate, changing the contact pressure or cushioning against the user’s head, thereby adapting the fit dynamically to reduce discomfort or pressure points.

[0039] The user access an outgoing microphone 111 installed with the frame 101 to issue voice commands to manually adjust the position of the rollers. The microphone 111 consists of a diaphragm, typically made of a thin, flexible material such as metal or plastic. When sound waves reach the microphone, leads to providing a vibrating movement to the diaphragm. These vibrations are directly proportional to the variations in air pressure caused by the sound waves. The diaphragm is coupled to a coil of wire, as the diaphragm vibrates, the coil moves within a magnetic field, inducing an electric current in the wire. This current is proportional to the amplitude and frequency of the sound waves. The electrical signal generated by the diaphragm-coil is transmitted to the microcontroller associated with the wearable glasses.

[0040] An array of directional microphones 106 is discreetly embedded around the front and sides of the frame 101, including within the upper rim and temple areas. This multi-directional placement is crucial to capture ambient sound from multiple directions, allowing the glasses to isolate the target speaker. Each microphone 111 picks up audio signals from its specific direction, and by combining these signals, the microcontroller determines the location of different sound sources.

[0041] The directional microphones 106 include voice activity detection sensors that activate only when speech is detected, conserving power and improving accuracy. These sensors are voice activity detection (VAD) sensors that work by continuously analysing incoming audio signals to distinguish between speech and non-speech sounds. They use protocols that detect characteristics typical of human speech such as specific frequency patterns, energy levels, and temporal dynamics and send a signal to the microcontroller when these speech features are present.

[0042] The array of microphones 106 is equipped with beam-forming and noise-cancellation protocols to filter out background noise and isolate the target speaker's voice in noisy environments. Beamforming is a signal processing protocol used to focus the microphone 111 array’s sensitivity in a specific direction, toward the target speaker. The array consists of directional microphones 106 spaced apart, and by adjusting the timing (phase) and amplitude of the signals received from each microphone, the microcontroller constructively combine sounds coming from the desired direction while destructively interfering with sounds from other directions. This creates a “beam” of enhanced sensitivity, effectively isolating the speaker’s voice from surrounding noise.

[0043] Noise-cancellation protocols reduce unwanted background sounds by identifying and subtracting noise components from the audio signal. This is often done to capture ambient noise and then applying protocols that analyze the noise characteristics and remove them from the main microphone 111 signals. The result is a cleaner audio output where the speaker’s voice is clearer and background noise is significantly minimized.

[0044] The glasses are configured with AI (artificial intelligence)-powered camera 110 to identify speakers and interpret hand signs or gestures. The camera 110 is integrated with a facial recognition and gesture recognition capabilities. The AI-powered camera 110 in the glasses captures real-time video of the user’s surroundings, and its integrated facial recognition and gesture recognition capabilities process this visual data to identify speakers and interpret hand signs or gestures. Facial recognition works by detecting and analysing key facial features such as the shape of the eyes, nose, mouth, and their spatial relationships using deep learning models trained on large datasets of faces.

[0045] Gesture recognition works by capturing video of hand or body movements using an embedded camera, then processing the images to isolate the relevant parts like hands through techniques such as background subtraction or color detection. The processor extracts key features such as finger positions and motion patterns, which are analysed by AI models trained to classify different gestures. Once recognized, each gesture is mapped to a specific command or meaning, enabling the glasses to interpret and respond to natural hand movements for intuitive user interaction. Together, these capabilities enable the glasses to understand who is speaking and what gestures are being made, providing contextual awareness and enhancing user interaction in real time.

[0046] The glasses also incorporate nanostructured capacitive sensors around the lens 103 edges to track eye movements of both the user and the speaker. The nanostructured capacitive sensors are placed around the edge of the lens 103 to track the eye movements of both the user and the person user is talking to. Nanostructured capacitive sensors around the lens 103 edges work by detecting changes in the electrical field caused by the movement of the eyes near the sensor surface. These sensors consist of tiny, nanoscale electrodes that form a capacitive field; when the eye or eyelid moves close to or across this field, it changes the capacitance value. By continuously measuring these variations, the sensors accurately track the direction and position of eye movements for both the user and the speaker, enabling precise monitoring of gaze and attention without interfering with vision.

[0047] This tracking facilitates accurate speaker identification and helps a translation module configured to provide real-time translation of spoken language, sign language, and gestures into a preferred language of the user. This module processes data from the directional microphones 106, AI camera, and sensors to deliver accurate and context-aware translations.

[0048] The translation module begins by capturing audio and visual data from its integrated sensors: the microphone 111 array records speech, the AI camera 110 captures facial expressions and hand gestures, and the capacitive eye-tracking sensors monitor where both the user and speaker are looking. The audio input is first processed through noise-cancellation and voice activity detection to isolate clear speech signals. Simultaneously, the AI camera 110 uses facial recognition to identify the speaker and gesture recognition protocol to interpret any sign language or hand motions. Eye-tracking data helps determine the speaker the user is focusing on, improving speaker identification and contextual relevance.

[0049] Next, the speech recognition engine converts the cleaned audio signals into text, while the gesture recognition protocol translates detected gestures or sign language into corresponding commands or words. These inputs are then fed into a natural language processing (NLP) protocol, which understands the meaning, context, and intent behind the spoken or gestured language. The NLP also resolves ambiguities and adapts the interpretation based on conversation context gathered from all inputs, including gaze and facial cues. Once the language content is fully processed and understood, the machine translation engine translates it into the user’s preferred language in real time.

[0050] The translated output is displayed through an augmented reality heads-up display (HUD) integrated directly into a pair electrochromic lens 103 provided with the frame 101, allowing the user to view translations without distraction. The augmented reality heads-up display (HUD) works by projecting the translated text or visuals directly onto the surface of the glasses’ lenses 103 within the user’s field of view. This projection uses tiny micro-displays or waveguide technology that bends light to overlay digital information seamlessly onto the real-world scene the user is looking at. The HUD aligns the translation output so it appears naturally positioned near the speaker without obstructing the user’s vision. This allows the user to read translations in real time while maintaining awareness of their surroundings, enabling distraction-free, intuitive communication.

[0051] The AI camera 110 features a motorized optical zoom and is linked to a global facial database to identify speakers and retrieve contextual data to optimize translation and interaction. The AI camera’s motorized optical zoom adjusts the lens 103 to zoom in or out smoothly, allowing it to focus closely on the speaker’s face for clearer image capture. This high-quality, zoomed-in image is then processed by facial recognition protocol that compare the captured facial features against a global facial database containing millions of known identities. Once the speaker is identified, the microcontroller retrieves the relevant contextual data from a cloud database such as the person’s name, language preferences, and communication style which helps optimize the translation accuracy and tailor interactions.

[0052] To improve audio clarity, the glasses include at least two ear units 107 mounted on articulated arms connected to the temple ends. These ear units 107 enhance sound quality by filtering background noise, ensuring that the user receives clear audio signals from the speaker or translation output.

[0053] Each ear unit 107 contains an amplifier that boosts the translated sound captured from the directional microphone array. The amplifier in each ear unit 107 works by receiving the audio signal captured by the directional microphone array and increasing its strength (volume) without significantly distorting the sound quality. The amplifier takes the weak electrical signals from the microphones 106 and uses integrated circuits to boost the signal’s amplitude. This amplified audio is then delivered clearly and loudly to the user’s ear through the speaker, ensuring the translated sound is easily heard even in noisy environments. The amplifier also helps maintain sound clarity and fidelity for better listening comfort.

[0054] A digital signal processor (DSP) is built into each ear unit 107 to filter out background noise and enhance important auditory cues, such as the speaker’s voice. This feature significantly improves the clarity of conversations between the user and the speaker, whether it be a celebrity, athlete, or politician. The digital signal processor (DSP) in each ear unit 107 works by receiving the incoming audio signals from the microphones 106 to analyse and modify these signals in real time.

[0055] The DSP filters out unwanted background noise by identifying and suppressing sounds that don’t match the characteristics of the speaker’s voice, such as static, crowd noise, or echoes. At the same time, the DSP enhances important auditory cues like speech frequencies and intonation patterns to make the speaker’s voice clearer and more distinct. This processing helps the user focus on the conversation by improving speech intelligibility, even in noisy or challenging environments, ensuring clearer communication with the speaker.

[0056] At regular intervals, the microcontroller suggests that the user take a hearing test if needed. If mild hearing loss is detected, the microcontroller slightly enhances the translated sound. In cases of severe hearing loss, the microcontroller significantly boosts the sound level to ensure they don't miss any important parts of conversations during major press events, interviews, or election coverage.

[0057] The outgoing microphone 111 captures the user's voice pitch and if the voice pitch deviates from a predefined range, the microcontroller provides haptic feedback alerts to guide the user in maintaining effective communication. The outgoing microphone 111 continuously captures the user’s voice by detecting the sound waves produced during speech and converting them into electrical signals. These signals are then analysed in real time by an audio processing protocol of the microcontroller to determine the pitch of the user’s voice. The microcontroller compares this detected pitch against a predefined optimal pitch range, which is set based on factors such as clarity, comfort, and effective communication standards.

[0058] If the user’s voice pitch deviates from this range, either becoming too high (which might sound strained or shrill) or too low (which might be difficult to hear or unclear), the microcontroller activates haptic feedback alerts, such as subtle vibrations delivered through the glasses. This tactile alert serves as an immediate, non-intrusive signal to the user, helping them become aware of their vocal modulation and encouraging them to adjust their pitch back within the ideal range.

[0059] The glasses are also equipped with a liquid crystal lens 108 housed in a concealed motorized vertical slider 109 above the frame 101. The crystal lens 108 move downward in front of the user's eyes to provide enlarged and sharper imaging of distant speakers, aiding better visual comprehension during interactions.

[0060] The motorized vertical slider 109 is mounted with the crystal lens 108 of the frame 101 that gently moves the liquid crystal lens 108 downward in front of the user's eyes. This allows the user to see a larger and sharper image of the celebrity, making it much easier to recognize faces even from a distance.

[0061] The motorized vertical slider 109 works by using a small, precise electric motor mounted within the glasses frame 101 to control the up-and-down movement of the liquid crystal lens 108. The microcontroller sends a signal to the motor gently slides the LC crystal lens 108 downward into the user’s line of sight. The liquid crystal lens 108 accordingly adjusts its optical properties, such as focus or magnification, electronically. This movement and adjustment allow the user to see a larger, clearer image of the speaker by effectively enhancing visual zoom and sharpness without requiring physical head movement or external magnifying tools.

[0062] The liquid crystal lens 108 works by using a layer of liquid crystal material sandwiched between transparent electrodes, which change the lens’s optical properties when an electric voltage is applied. In its default state, the liquid crystals allow light to pass through normally, but when voltage is introduced, the crystals realign to alter the lens’s refractive index, bending light differently across the surface. This change creates a lensing effect that adjust focus or magnification without any moving parts.

[0063] The frame 101 is embedded with an optical sensor that detects when the glasses are not in use. The optical sensor works by detecting changes in light levels to determine whether the glasses are being worn. When the glasses are in use, the sensor, usually positioned near the bridge 104, the sensor detects the presence of the user's face by sensing reflected infrared light or ambient light patterns. If the sensor no longer detects these cues, it determines that the glasses are not in use.

[0064] At the time when frame 101 is not in use, a flap 201 curved-shaped cover flap 201 attached to the outer rims 102 is closed for preventing damage and contamination. This flap 201 acts as a shield to prevent dust, dirt, fingerprints, or physical impacts from damaging sensitive areas such as the lenses, sensors, cameras, or displays. Designed to fit the contour of the frame 101, the cover creates a sealed barrier that helps maintain the cleanliness and integrity while in storage or transit. The flap 201 may be made from durable, lightweight materials like polycarbonate or carbon fiber to avoid scratching the lenses.

[0065] The microcontroller sends a signal to motorized hinge joints 202 connecting the flap 201 to the rims 102 to deploy the flap 201. The motorized hinge joints 202 used herein, is a piece of metal that joins two sides or items together and allows it to be opened or closed by revolving along the longitudinal axis whose operation is governed by a DC motor to deploy the flap 201 over the lenses.

[0066] Environmental awareness and user safety are enhanced by integrating a GPS module with the microcontroller. The GPS tracks the user’s real-time location, and the AI camera 110 analyzes surroundings upon arrival at specific locations. The GPS (Global Positioning System) module works in sync with a magnetometer provides enhanced positioning and orientation information of the location of the user.

[0067] The GPS module receives signals from multiple satellites in orbit around the Earth. These satellites transmit precise timing and position information of the user. The GPS module receives these signals and uses the time delay between transmission and reception to calculate the distance between the GPS module and each satellite. By triangulating the distances from multiple satellites, the GPS module determines its own position on the Earth's surface. This position is typically given in latitude and longitude coordinates. The magnetometer measures the strength and direction of the magnetic field in its vicinity. The magnetometer detects the Earth's magnetic field, which is approximately aligned with the Earth's geographic north-south axis. By utilizing the magnetometer's measurements, the GPS module determine the band heading or orientation relative to magnetic north. The magnetometer provides information about the direction of the Earth's magnetic field, which is compared with the band position information obtained from the GPS module. The outputs of the GPS module and the magnetometer are combined and processed by the microcontroller in order to determine the real-time location of the user. The user’s location data is then cross-referenced with a cloud-based AI database that monitor environmental hazards (like fires, accidents, or disasters) in real time.

[0068] An audio alert is generated via the ear units 107 to notify the user of the environmental events such as fire outbreaks or accidents. To achieve this, the ear units 107 use distinct tones including sharp beeps or rising pitch alarms for different types of events. For alerting in more hazardous situations, voice prompts are used to provide detailed information, offering not just an alert but also actionable guidance.

[0069] To maintain lens 103 clarity, a dust sensor continuously monitors the lens 103 surfaces. The sensor emits a small beam of light at the lens 103 and then measures the light that is scattered or reflected back. When dust or debris is present, it interferes with the light path, causing changes in the reflected signal. These changes are analysed by the microcontroller and compared to a pre-fed database to determine the amount and location of dust accumulation.

[0070] When dust or debris is detected, the microcontroller activates a series of micro air-jet nozzles positioned along the inner and outer upper frame 101, which blow away any particles, ensuring clear vision. The micro air-jet nozzles work by using a miniature compressor integrated within the glasses frame 101. When the dust sensor detects particles on the lens, it signals the microcontroller, which then send a signal to a motor connected to these nozzles. The nozzles, strategically placed along the inner and outer upper frame 101, release precise bursts of compressed air directed across the lens 103 surface. These controlled air jets dislodge and blow away dust, debris, or small particles without the need for physical wiping, ensuring that the lens 103 remains clean and vision stays clear.

[0071] Lastly, a user-interface inbuilt within a computing unit allows the user to manage various functions such as event schedules, maps, translation preferences, and hearing or vision profiles. These settings are securely stored in a connected database, providing a personalized and adaptable user experience. The user either selects from a list of options provided on the display or manually enters the details, where the user is required to enter details for event schedules, maps, translation preferences, and hearing or vision profiles.

[0072] The computing unit is linked with a microcontroller embedded within the housing via an integrated communication module that includes but is not limited to a GSM (Global System for Mobile Communication) module which is capable of establishing a wireless network between the microcontroller and the computing unit to assist the user in managing the functions of the glasses remotely.

[0073] A battery (not shown in figure) is associated with the glasses to supply power to electrically powered components which are employed herein. The battery is comprised of a pair of electrodes named as a cathode and an anode. The battery uses a chemical reaction of oxidation/reduction to do work on charge and produce a voltage between their anode and cathode and thus produces electrical energy that is used to do work in the glasses.

[0074] The present invention works best in the following manner, where the transparent frame 101 as disclosed in the invention is embedded with advanced electronics to enhance user interaction, communication, and safety, inked with the microcontroller that governs various interconnected components. The frame 101 houses electrochromic lenses 103 that auto-adjust tint in response to light detected by photodiodes. Facial and eye tracking are enabled through optical and nanostructured capacitive sensors, while the motorized nose bridge 104 and temple arms 105 with DC motor-driven ball-and-socket joints adapt to the user’s facial structure. Piezo-resistive sensors detect strain, prompting motorized foam rollers for comfort. The AI-powered camera 110 with facial and gesture recognition, motorized optical zoom, and access to the global facial database identifies speakers and interprets gestures. Directional microphones 106, equipped with beamforming and noise-cancellation protocols, work with voice activity detection sensors to isolate speech, while digital signal processors and amplifiers in ear units 107 enhance audio clarity. The translation module integrates data from sensors and the AI camera 110 to convert speech, gestures, and sign language into the user’s preferred language, displayed via the augmented reality heads-up display. The GPS module combined with the magnetometer provides real-time location and orientation, allowing cloud-synced environmental hazard detection. Micro air-jet nozzles, triggered by dust sensors, maintain lens 103 clarity. The motorized vertical slider 109 deploys the liquid crystal lens 108 for visual magnification. The computing unit with the GSM module connects wirelessly to the microcontroller, enabling remote management of maps, schedules, translation preferences, and sensory profiles, all securely stored in the connected database for the personalized, adaptive experience.

[0075] Although the field of the invention has been described herein with limited reference to specific embodiments, this description is not meant to be construed in a limiting sense. Various modifications of the disclosed embodiments, as well as alternate embodiments of the invention, will become apparent to persons skilled in the art upon reference to the description of the invention. , Claims:1) A multimodal language translation and communication wearable glasses, comprising:

i) a transparent frame 101 developed to be worn by a user and adapted to rest on the user’s nose and ears;
ii) an array of directional microphones configured to capture ambient sound from multiple directions;
iii) an AI (artificial intelligence)-powered camera 110 integrated with facial recognition and gesture recognition capabilities;
iv) an augmented reality heads-up display (HUD) integrated into a pair of electrochromic lens 103 provided with the frame 101;
v) a translation module configured to provide real-time translation of spoken language, sign language, and gestures into a preferred language of the user;
vi) at least two ear units 107 mounted on an articulated arm connected to a pair of telescopic temple arms 105 provided with the frame 101 to enhance audio clarity and filter background noise; and
vii) a microcontroller configured to coordinate data processing from the microphones 106, AI camera, sensors, translation module, and AR HUD,
wherein the microcontroller facilitates real-time context-aware language translation and communication assistance for user interacting with speakers of different languages or communication modes.

2) The wearable glasses as claimed in claim 1, wherein the frame 101 is constructed with a pair of lens holding rims, a motorized nose bridge 104, the rims 102 holding the electrochromic lens 103, and telescopic temple arms 105 with cushioning and pressure sensors that rest over the ears of the user.

3) The wearable glasses as claimed in claim 1, wherein the directional microphone array is paired with beam-forming and noise-cancellation protocols to isolate target speaker voices in noisy environments and includes a voice activity detection sensor which activate only during speech.

4) The wearable glasses as claimed in claim 1, wherein an outgoing microphone 111 is configured to capture the user’s voice pitch and provide haptic feedback alerts if the pitch is outside of a predefined range.

5) The wearable glasses as claimed in claim 1, wherein a liquid crystal lens 108 is housed in a concealed motorized vertical slider 109 above the frame 101, configured to move downward in front of the user’s eyes for enlarged and sharper imaging of the distant speakers.

6) The wearable glasses as claimed in claim 1, wherein a curved-shaped cover flap 201 is attached to the outer rims 102 of the lenses 103 via motorized hinge joints 202, configured to automatically close and protect the lenses 103, based on an optical sensor detecting the unused state of the frame.

7) The wearable glasses as claimed in claim 1, wherein nanostructured capacitive sensors are provided around the lens 103 edges to track eye movements of both the user and speaker, enabling accurate speaker identification and translation focus.

8) The wearable glasses as claimed in claim 1, wherein a GPS (Global Positioning System) module is integrated with the microcontroller tracks real-time reporting location, and the camera 110 analyze the surroundings upon arrival at a reporting location, detecting environmental events such as fire outbreaks or accidents, and communicates alerts to the user via the ear units 107.

9) The wearable glasses as claimed in claim 1, wherein a dust sensor is integrated to monitor lens 103 surfaces and triggers a plurality of micro air-jet nozzles positioned along the inner and outer upper frame 101 to blow away dust or debris.

10) The wearable glasses as claimed in claim 1, wherein a user-interface is inbuilt in a computing unit, allowing user(s) to manage event schedules, maps, translation preferences, and hearing or vision profiles securely stored in a connected database.

Documents

Application Documents

# Name Date
1 202521056016-STATEMENT OF UNDERTAKING (FORM 3) [10-06-2025(online)].pdf 2025-06-10
2 202521056016-REQUEST FOR EXAMINATION (FORM-18) [10-06-2025(online)].pdf 2025-06-10
3 202521056016-REQUEST FOR EARLY PUBLICATION(FORM-9) [10-06-2025(online)].pdf 2025-06-10
4 202521056016-PROOF OF RIGHT [10-06-2025(online)].pdf 2025-06-10
5 202521056016-POWER OF AUTHORITY [10-06-2025(online)].pdf 2025-06-10
6 202521056016-FORM-9 [10-06-2025(online)].pdf 2025-06-10
7 202521056016-FORM FOR SMALL ENTITY(FORM-28) [10-06-2025(online)].pdf 2025-06-10
8 202521056016-FORM 18 [10-06-2025(online)].pdf 2025-06-10
9 202521056016-FORM 1 [10-06-2025(online)].pdf 2025-06-10
10 202521056016-FIGURE OF ABSTRACT [10-06-2025(online)].pdf 2025-06-10
11 202521056016-EVIDENCE FOR REGISTRATION UNDER SSI(FORM-28) [10-06-2025(online)].pdf 2025-06-10
12 202521056016-EVIDENCE FOR REGISTRATION UNDER SSI [10-06-2025(online)].pdf 2025-06-10
13 202521056016-EDUCATIONAL INSTITUTION(S) [10-06-2025(online)].pdf 2025-06-10
14 202521056016-DRAWINGS [10-06-2025(online)].pdf 2025-06-10
15 202521056016-DECLARATION OF INVENTORSHIP (FORM 5) [10-06-2025(online)].pdf 2025-06-10
16 202521056016-COMPLETE SPECIFICATION [10-06-2025(online)].pdf 2025-06-10
17 Abstract.jpg 2025-06-25
18 202521056016-FORM-26 [24-10-2025(online)].pdf 2025-10-24