Sign In to Follow Application
View All Documents & Correspondence

Speech Enhancement Device

Abstract: A speech enhancement device, comprising a housing 101 with a display unit 102 attached, which interactively shows words, phrases, storylines, and translations in a child’s native language to aid pronunciation training, a microphone unit 103 monitors the child’s speech and provides real-time feedback on inaccuracies, a speaker unit 104 emits controlled background noises and rhymes to simulate real-world environments and engage the child in varied auditory conditions, an AI-powered camera 105 mounted externally tracks mouth movements and facial articulation to provide immediate corrective feedback, a speech therapy tool 106 features a spoon-like structure 106a with a bowl 106b inserted via a hydraulic piston 106c, having an expandable ring arrangement 107 of curved rods 107a connected by motorized hinges 107b and covered with medical-grade fabric for comfort and safety, the tool 106 also includes a vibrating unit 106d with intensity adjusted in real time based on therapeutic needs.

Get Free WhatsApp Updates!
Notices, Deadlines & Correspondence

Patent Information

Application #
Filing Date
13 August 2025
Publication Number
35/2025
Publication Type
INA
Invention Field
ELECTRONICS
Status
Email
Parent Application

Applicants

SR University
Ananthasagar, Hasanparthy (PO), Warangal-506371, Telangana, India.

Inventors

1. Suresh Kumar Mandala
SR University, Ananthasagar, Hasanparthy (PO), Warangal-506371, Telangana, India.
2. Neelima Gurrapu
SR University, Ananthasagar, Hasanparthy (PO), Warangal-506371, Telangana, India.
3. Jakku Harshavardhan
SR University, Ananthasagar, Hasanparthy (PO), Warangal-506371, Telangana, India.
4. Kandhagatla Shreya
SR University, Ananthasagar, Hasanparthy (PO), Warangal-506371, Telangana, India.
5. Palle Akshith Reddy
SR University, Ananthasagar, Hasanparthy (PO), Warangal-506371, Telangana, India.
6. Golla Chitraleka
SR University, Ananthasagar, Hasanparthy (PO), Warangal-506371, Telangana, India.
7. Kusam Deepak Raj
SR University, Ananthasagar, Hasanparthy (PO), Warangal-506371, Telangana, India.
8. Chandupatla Vyshnavi
SR University, Ananthasagar, Hasanparthy (PO), Warangal-506371, Telangana, India.

Specification

Description:FIELD OF THE INVENTION

[0001] The present invention relates to a speech enhancement device that is capable of assisting children in improving pronunciation, articulation, and language skills through interactive learning, real-time feedback, and adaptive exercises tailored to individual progress and therapeutic requirements.

BACKGROUND OF THE INVENTION

[0002] Speech enhancement is crucial for improving clarity, accuracy, and confidence in verbal communication, particularly for children with speech difficulties. The speech enhancement aids in correcting pronunciation, breathing, and articulation, facilitating effective learning and social interaction. Early intervention supports language development and helps prevent long-term communication issues. However, speech enhancement faces challenges such as background noise interference, inconsistent speech patterns, and difficulty in accurately identifying pronunciation errors. Additionally, children’s varying speech abilities and limited attention spans make providing real-time feedback more complex. Delivering personalized, engaging training while simultaneously monitoring breathing and articulation further increases the complexity of effective speech improvement.

[0003] Traditionally, speech enhancement relies on manual tools like flashcards, audio recordings, and face-to-face therapy sessions with speech therapists. Some electronic devices offer basic amplification or noise reduction but lack interactive feedback or personalized training. These traditional methods often require constant human supervision and do not automatically adapt to the learner’s progress. Automation is limited or absent, making real-time correction and customization difficult. As a result, progress become slow, inconsistent, and dependent on the availability of trained professionals. The lack of integrated technology also means these devices not track performance data or provide dynamic, engaging learning experiences tailored to individual needs.

[0004] US7191127B2 discloses about a method and apparatus for reducing noise in a speech signal. A handset or remote unit provides to users with a hearing deficiency, a first mode of operation where noise suppressant/speech enhancement algorithms are used during any auditory-related service. There is also provided, in a related mode of operation, speech filtering for reducing noise in a speech signal received through the microphone and outputting the filtered sound to the speaker. The handset includes a microphone for receiving an auditory sound, a receiver for receiving an auditory signal and a speech filter for suppressing noise in the auditory signal and sound. The speech filter also may be configured to shift the frequency and/or alter the intensity of the auditory signal and sound. The speaker is used for amplifying and outputting the enhanced speech component as an audible sound.

[0005] EP4371112A1 discloses about a method for enhancing audio signals is provided. In some implementations, the method involves (a) obtaining a training set comprising a plurality of training samples, each training sample comprising a distorted audio signal and a clean audio signal. In some implementations, the method involves (b), for a training sample of the plurality of training samples: obtaining a frequency-domain representation of the distorted audio signal; providing the frequency-domain representation to a convolutional neural network (CNN) comprising a plurality of convolutional layers and to a recurrent element, wherein an output of the recurrent element is provided to a subset of the plurality of convolutional layers; generating a predicted enhancement mask, wherein the CNN generates the predicted enhancement mask; generating a predicted enhanced audio signal based on the predicted enhancement mask; and updating weights associated with the CNN and the recurrent element based on the predicted enhanced audio signal.

[0006] Conventionally, many devices are available in market that assist users in speech enhancement. However, these existing devices lack in providing real-time personalized feedback, adaptive training, and integration of multiple sensory inputs. They often fail to track progress or customize sessions, limiting effectiveness and engagement, especially for children with varied speech difficulties.

[0007] In order to overcome the aforementioned drawbacks, there exists a need in the art to develop a device that requires to be capable of providing real-time, personalized speech training by integrating auditory, visual, and tactile feedback. In addition, the device also autonomously adapt to user’s progress, monitor speech and articulation, and generate performance reports without requiring constant human supervision.

OBJECTS OF THE INVENTION

[0008] The principal object of the present invention is to overcome the disadvantages of the prior art.

[0009] An object of the present invention is to develop a device that is capable assisting users in developing clear pronunciation and better articulation by providing structured, interactive speech training sessions designed to improve their language and communication skills.

[0010] Another object of the present invention is to develop a device that is capable of delivering real-time feedback during speech practice, allowing immediate correction of mistakes and helping children learn proper pronunciation through repetition, guidance, and continuous improvement.

[0011] Another object of the present invention is to develop a device that is capable of adjusting training content and difficulty levels autonomously based on each child’s individual progress for ensuring personalized learning experiences that keep the child engaged and motivated.

[0012] Yet another object of the present invention is to develop a device that is capable of tracking and evaluating child’s speech development over time, generating regular performance summaries that is reviewed by guardians, educators, or therapists for effective monitoring.

[0013] The foregoing and other objects, features, and advantages of the present invention will become readily apparent upon further review of the following detailed description of the preferred embodiment as illustrated in the accompanying drawings.

SUMMARY OF THE INVENTION

[0014] The present invention relates to a speech enhancement device that is capable of providing interactive and personalized training for children by offering real-time corrective feedback on pronunciation and articulation, adapts to the child's progress, and incorporates breathing exercises and specialized means to facilitate comprehensive speech development in varied auditory settings.

[0015] According to an embodiment of the present invention, a speech enhancement device, comprising a housing, a display unit attached to the housing, adapted to interactively display words, phrases, storylines, and translations in a child’s native language for pronunciation training, a microphone unit installed on the housing and synchronized with a microcontroller, adapted to monitor the child's speech and provide real-time feedback based on detected inaccuracies, a speaker unit provided with the housing to emit controlled background noises and rhymes to simulate real-world environments and engage the child in pronunciation practice under varied auditory conditions, a thermal flow sensor integrated within the housing to detect airflow and monitor the child’s breathing patterns during speech, irregularities in breathing patterns triggers guided breathing exercises over the display unit, an AI-powered camera mounted on exterior of the housing, configured to track mouth movements and facial articulation of the child to provide real-time corrective feedback, a speech therapy tool comprising a spoon-like structure with a bowl mounted on the housing, inserted into the child’s mouth via a hydraulic piston, the bowl having an expandable ring arrangement for articulatory exercise, the microcontroller utilizes an IoT-enabled cloud database adapted to store child-specific data including name, age, class, and practice level, to generate customized training sessions and weekly progress reports accessible to authorized users.

[0016] According to another embodiment of the present invention, the device further comprises of the expandable ring arrangement, including a plurality of curved rods connected via motorized hinges and integrated with medical-grade fabric for safety and comfort during use, the speech therapy tool includes a vibrating unit with intensity dynamically adjusted in real-time based on therapeutic requirements, the microcontroller is trained to analyze the child’s pronunciation accuracy over time and modify the vocabulary and exercise complexity accordingly to sustain engagement and improvement, the microphone unit captures speech for NLP-based pronunciation comparison and generates immediate audio-visual feedback for speech correction, the cloud database stores weekly performance reports summarizing vocabulary learned, quiz accuracy, and areas requiring improvement, and transmits the report to authorized guardians or teachers, the microcontroller is adapted to dynamically adjust training difficulty based on user progress, analyze performance data, and issue personalized feedback, the microcontroller displays pitch modulation exercises and dynamic tongue twisters with gradually increasing difficulty levels to improve articulation and control under pressure.

[0017] While the invention has been described and shown with particular reference to the preferred embodiment, it will be apparent that variations might be possible that would fall within the scope of the present invention.

BRIEF DESCRIPTION OF THE DRAWINGS

[0018] These and other features, aspects, and advantages of the present invention will become better understood with regard to the following description, appended claims, and accompanying drawings where:
Figure 1 illustrates an isometric view of a speech enhancement device.

DETAILED DESCRIPTION OF THE INVENTION

[0019] The following description includes the preferred best mode of one embodiment of the present invention. It will be clear from this description of the invention that the invention is not limited to these illustrated embodiments but that the invention also includes a variety of modifications and embodiments thereto. Therefore, the present description should be seen as illustrative and not limiting. While the invention is susceptible to various modifications and alternative constructions, it should be understood, that there is no intention to limit the invention to the specific form disclosed, but, on the contrary, the invention is to cover all modifications, alternative constructions, and equivalents falling within the spirit and scope of the invention as defined in the claims.

[0020] In any embodiment described herein, the open-ended terms "comprising," "comprises,” and the like (which are synonymous with "including," "having” and "characterized by") may be replaced by the respective partially closed phrases "consisting essentially of," consists essentially of," and the like or the respective closed phrases "consisting of," "consists of, the like.

[0021] As used herein, the singular forms “a,” “an,” and “the” designate both the singular and the plural, unless expressly stated to designate the singular only.

[0022] The present invention relates to a speech enhancement device that is capable of improving pronunciation, articulation, and language skills in children by offering interactive learning, real-time corrective feedback, progress tracking, and personalized training sessions based on individual performance and developmental needs.

[0023] Referring to Figure 1, an isometric view of a speech enhancement device is illustrated, comprising a housing 101 positioned on a surface, a display unit 102 attached to the housing 101, a microphone unit 103 installed on the housing 101, a speaker unit 104 provided with the housing 101, an AI-powered camera 105 mounted on exterior of the housing 101, a speech therapy tool 106 comprising a spoon-like structure 106a with a bowl 106b mounted on the housing 101 via a hydraulic piston 106c, a vibrating unit 106d and the bowl 106b having an expandable ring arrangement 107 includes a plurality of curved rods 107a connected via motorized hinges 107b.

[0024] The device disclosed herein includes a housing 101 is developed to be positioned on a flat surface. The housing 101 herein is cuboidal in shape and includes all necessary component of the device for assisting users in speaks enhancement.

[0025] A push button is equipped with the device for activating and deactivating the device. The push button is accessed by the user for activating the device. When the user presses the push button, the electrical circuit is completed, which in response turns the device on. The push button is integrated with an actuator and a spring, which are automatically activated when pressed. They work together to move the internal contact, completing the circuit and allowing electrical current to flow, thereby activating the device.

[0026] When the push button is pressed, the button sends a signal (usually a change in voltage or current) to an inbuilt microcontroller associated with the device to either power up or shut down the microcontroller. Conversely, releasing the button allows the spring to return to its original position, breaking the circuit and sending the signal to deactivate the device. The microcontroller is pre-fed to detect this signal and respond accordingly. The microcontroller used herein is pre-fed using artificial intelligence and machine learning protocols to coordinate the working of the device.

[0027] A display unit 102 is mounted on the housing 101 and is configured to interactively present words, phrases, storylines, and translations in a child’s native language to facilitate pronunciation training. In a preferred embodiment, the display unit 102 is a touch-enabled electronic visual interface, also referred to as a touchscreen, designed to detect and respond to the presence and location of a touch within the display area.

[0028] The touchscreen typically comprises several layers, including a display panel, a touch sensor, and a controller. The outermost layer is the display panel, which is a Liquid Crystal Display (LCD) responsible for rendering visual content. Positioned beneath this is a transparent touch sensor, commonly made of indium tin oxide (ITO), configured as a grid of intersecting conductive lines forming multiple touch points. When the user touches the display, the touch sensor detects changes in electrical signals at the point of contact. These signals are transmitted to a dedicated touch controller, which filters noise and determines the precise location of the touch.

[0029] The touch data is then sent to the microcontroller, which serves as the central processing unit for managing device operations. The microcontroller is connected to an IoT-enabled cloud database that stores child-specific data, including name, age, class, and practice level. Based on this data, the microcontroller generates customized pronunciation training sessions and displays age-appropriate content such as vocabulary, phrases, interactive storylines, and translations directly on the screen. The child engages with the content by tapping words or icons, which triggers audio playback and pronunciation prompts.

[0030] Additionally, the display unit 102 is configured to conduct randomized quizzes using previously learned vocabulary. In the event of incorrect responses, the display provides immediate visual feedback, including the correct pronunciation and meaning, to reinforce retention and learning. As the child interacts with the display unit 102, performance metrics—such as quiz scores and completed activities—are recorded and sent to the cloud database. This enables the generation of weekly progress reports, which are accessible to authorized guardians, educators, or therapists. Overall, the feedback-driven and adaptive visual interface ensures a personalized and engaging learning experience tailored to each child’s developmental and linguistic needs.

[0031] In an embodiment of the present invention, a communication module is linked with the microcontroller and activated for establishing a wireless connection between the microcontroller and a computing unit (includes, but not limited to smartphone, tablet or laptop) and inbuilt with a user-interface that is accessed by the authorized guardians, educators, or therapists. The communication module used herein includes, but not limited to Wi-Fi (Wireless Fidelity) module, Bluetooth module, GSM (Global System for Mobile Communication) module. The communication module used herein is preferably a Wi-Fi module that is a hardware component that enables the microcontroller to connect wirelessly with the computing unit. The Wi-Fi module works by utilizing radio waves to transmit and receive data over short distances. The core functionality relies on the IEEE 802.11 standards, which define the protocols for wireless local area networking (WLAN). Once connected, the module allows the microcontroller to send and receive data through data packets.

[0032] A microphone unit 103 is mounted on the housing 101 and is configured to capture and monitor the child’s speech in real time, providing immediate feedback based on detected pronunciation inaccuracies. The microphone unit 103 typically comprises a diaphragm attached to a moving coil situated within a magnetic field. When sound waves produced by the child reach the diaphragm, they cause the coil to vibrate. These vibrations induce movement of the coil within the magnetic field, thereby generating a corresponding electrical signal.

[0033] This analog audio signal is transmitted to the microcontroller. The microcontroller digitizes the signal and processes it using embedded speech recognition protocols and natural language processing (NLP) techniques. The processed data is then compared against predefined pronunciation models or expected phonetic patterns. In the event of a mismatch or deviation from correct pronunciation, the microcontroller triggers immediate audio-visual feedback via the display unit 102 and a speaker unit 104 is installed on the housing 101 to guide the child toward correct articulation. Additionally, the microphone unit 103 utilizes noise-cancellation filters or gain control means to ensure accurate speech capture even in environments with moderate background noise. This continuous feedback loop helps the child improve pronunciation accuracy through real-time correction and repetition.

[0034] The speaker unit 104 is configured to emit controlled background sounds, rhymes, and rhythmic audio cues to simulate real-world auditory environments. This functionality is designed to engage the child in pronunciation practice under varied listening conditions, enhancing auditory discrimination and articulation skills.

[0035] The speaker unit 104 operates by converting electrical signals into audible sound waves. Structurally, it consists of a diaphragm—also referred to as a cone—attached to a coil of wire known as the voice coil, which is positioned within a magnetic field created by surrounding permanent magnets. When an electrical signal, generated by the microcontroller or received from other components, passes through the voice coil, it creates a varying magnetic field that interacts with the magnetic field of the permanent magnets. This interaction causes the voice coil and the attached diaphragm to move back and forth.

[0036] As the diaphragm vibrates, it pushes and pulls the surrounding air, producing sound waves that replicate the electrical signal's characteristics. These sound waves are then emitted as speech cues, background noise, or instructional audio, depending on the training session. The speaker unit 104 is also work in coordination with the display unit 102 to deliver synchronized audio-visual content, reinforcing the child’s pronunciation practice. By simulating diverse auditory environments, the speaker trains the child to articulate correctly in different sound conditions, thereby improving speech clarity and comprehension in real-world settings.

[0037] A thermal flow sensor is integrated within the housing 101 and is configured to detect airflow and monitor the child’s breathing patterns during speech. The thermal flow sensor typically consists of a heating element and temperature sensors positioned on either side of the airflow path. The heating element warms the surrounding air, and as the child breathes, airflow causes heat to be carried away from the sensor in proportion to the rate and direction of the breath. The temperature sensors measure these variations in heat distribution, allowing the device to calculate the velocity and pattern of airflow. By continuously monitoring these airflow changes, the sensor detects irregularities in the child’s breathing patterns, such as shallow breaths, pauses, or inconsistent respiratory rhythms. Upon identifying such irregularities, the microcontroller triggers guided breathing exercises that are displayed on the display unit 102, providing visual prompts and instructions to help the child regulate their breathing.

[0038] An AI-powered camera 105 is mounted on the exterior of the housing 101 and is configured to monitor the child’s mouth movements and facial articulations in real time, providing immediate corrective feedback to enhance speech accuracy. The camera 105 comprises an image-capturing module, including a set of lenses that capture multiple images of the child’s facial region from various angles. These captured images are stored temporarily in the camera’s memory as optical data.

[0039] The camera 105 further includes an onboard processor embedded with artificial intelligence protocols designed to analyze the optical data. This processor processes the images to extract relevant features, such as lip positioning, tongue movement, and jaw articulation. The extracted visual data is then converted into digital signals in the form of pulses and bits, which are transmitted to the microcontroller.

[0040] Upon receiving the processed data, the microcontroller evaluates the child’s articulation patterns against predefined models of correct speech movements. Based on this analysis, the microcontroller provides real-time corrective feedback through the display unit 102 and speaker unit 104 for guiding the child toward improved pronunciation and facial articulation during speech exercises.

[0041] A speech therapy tool 106 comprising a spoon-like structure 106a with a bowl 106b is mounted on the housing 101, develop to be inserted into the child’s mouth via a hydraulic piston 106c. The hydraulic piston 106c operates by regulating the flow of hydraulic fluid into its cylinder, which controls the extension and retraction of an elongated actuator rod connected to the bowl 106b. When hydraulic fluid is pumped into the piston 106c chamber, pressure is generated that pushes the actuator rod outward, extending the bowl 106b into the child’s oral cavity for therapeutic exercise.

[0042] To retract the bowl 106b, the hydraulic fluid is released from the piston chamber, allowing the actuator rod to withdraw smoothly back into the cylinder. The position and movement of the piston are precisely controlled by valves that manage fluid flow, ensuring accurate, stable, and smooth extension and retraction of the bowl 106b during use. The bowl 106b is securely attached to the actuator rod to maintain stability throughout the motion, while the hydraulic piston 106c provides the necessary force and fine control required for correct positioning and alignment within the mouth. This arrangement facilitates targeted articulatory exercises by enabling controlled manipulation of oral muscles through an expandable ring arrangement 107 within the bowl 106b.

[0043] The expandable ring arrangement 107 comprises a plurality of curved rods 107a interconnected through motorized hinges 107b, allowing controlled expansion and contraction. These rods 107a are encased in medical-grade, biocompatible fabric to ensure safety, hygiene, and user comfort during oral articulation exercises. Each motorized hinges 107b includes a pair of hinge leaves affixed to the adjoining surfaces of adjacent curved rods 107a. These leaves are connected via a cylindrical member that houses a rotating shaft coupled to a miniature DC (Direct Current) motor. The DC motor provides the necessary torque to control the rotational movement of the shaft. When the shaft rotates in the clockwise direction, the hinges 107b opens, causing the connected rods 107a to move apart and expand the ring structure. Conversely, counterclockwise rotation results in hinges 107b closure, contracting the ring assembly.

[0044] The microcontroller precisely controls the actuation of the DC motors based on pre-fed articulation exercises or real-time therapeutic feedback. This allows the expandable ring arrangement 107 to perform targeted movements, providing resistance or support to the child’s oral muscles, thereby assisting in strengthening and improving articulation accuracy during speech therapy sessions.

[0045] The speech therapy tool 106 further includes a vibrating unit 106d, wherein the vibration intensity is dynamically adjusted in real time based on the therapeutic requirements of the child. The vibrating unit 106d integrated into the speech therapy tool 106 functions to stimulate the oral muscles and enhance sensory feedback during articulation exercises. The vibration unit typically consists of a small electric motor with an offset weight attached to its shaft. When the motor rotates, the imbalance causes rapid oscillations, generating vibrations that are transmitted through the spoon-like structure 106a to the oral cavity. The intensity of these vibrations is dynamically controlled in real time by the microcontroller, which adjusts the motor’s speed based on the therapeutic needs of the child. These requirements determined through pre-fed therapy protocols, real-time feedback from the sensors monitoring the child’s performance, or data retrieved from the cloud database storing the child’s progress history. The microcontroller is trained to continuously analyze the child’s pronunciation accuracy over time using embedded speech recognition and natural language processing protocols. For instance, higher vibration levels applied during specific exercises requiring muscle stimulation or attention, while gentler vibrations are used for relaxation or sensitivity training.

[0046] The microcontroller also generates and displays pitch modulation exercises and dynamic tongue twisters, progressively increasing in difficulty to enhance articulation, breath control, and clarity under pressure. The speaker unit 104 decomposes complex words into individual syllables, enabling repeat-after-me style learning to improve phonetic articulation and speech clarity. This ensures that the therapy remains both personalized and responsive, delivering an optimized speech development experience tailored to the individual needs of each user.

[0047] The present invention works best in the following manner, where the housing 101 as disclosed in the invention supports the display unit 102 is configured to interactively present words, phrases, storylines, and translations in the child’s native language using the touch-sensitive LCD panel controlled by the microcontroller. The microphone unit 103 captures speech input, which is processed by the microcontroller using natural language processing protocols to identify pronunciation inaccuracies and generate real-time audio-visual feedback. The speaker unit 104 emits controlled background noises and rhymes to simulate real-world auditory conditions for more immersive training. The thermal flow sensor monitors the child’s breathing patterns during speech, and if irregularities are detected, the display unit 102 triggers guided breathing exercises. The AI-powered camera 105 tracks mouth movements and facial articulation to provide real-time corrective visual feedback. The speech therapy tool 106 includes the spoon-like structure 106a with the bowl 106b, actuated by the hydraulic piston 106c, and features the expandable ring arrangement 107 formed of curved rods 107a with motorized hinges 107b enclosed in medical-grade fabric. The tool 106 also integrates the vibrating unit 106d with dynamically adjustable intensity. The microcontroller utilizes the IoT-enabled cloud database to store child-specific data, generate personalized training sessions, adjust vocabulary and exercise complexity over time, and produce weekly performance reports for authorized guardians. The device further includes pitch modulation drills, tongue twisters, and syllable-based speaker guidance to enhance phonetic learning through repeat-after-me interaction.

[0048] Although the field of the invention has been described herein with limited reference to specific embodiments, this description is not meant to be construed in a limiting sense. Various modifications of the disclosed embodiments, as well as alternate embodiments of the invention, will become apparent to persons skilled in the art upon reference to the description of the invention. , C , Claims:1) A speech enhancement device, comprising:

i) a housing 101;
ii) a display unit 102 attached to the housing 101, adapted to interactively display words, phrases, storylines, and translations in a child’s native language for pronunciation training;
iii) a microphone unit 103 installed on the housing 101 and synchronized with a microcontroller, adapted to monitor the child's speech and provide real-time feedback based on detected inaccuracies;
iv) a speaker unit 104 provided with the housing 101 to emit controlled background noises and rhymes to simulate real-world environments and engage the child in pronunciation practice under varied auditory conditions;
v) a thermal flow sensor integrated within the housing 101 to detect airflow and monitor the child’s breathing patterns during speech, irregularities in breathing patterns triggers guided breathing exercises over the display unit 102;
vi) an AI-powered camera 105 mounted on exterior of the housing 101, configured to track mouth movements and facial articulation of the child to provide real-time corrective feedback; and
vii) a speech therapy tool 106 comprising a spoon-like structure 106a with a bowl 106b mounted on the housing 101, inserted into the child’s mouth via a hydraulic piston 106c, the bowl 106b having an expandable ring arrangement 107 for articulatory exercise.

2) The device as claimed in claim 1, wherein the microcontroller utilizes an IoT-enabled cloud database adapted to store child-specific data including name, age, class, and practice level, to generate customized training sessions and weekly progress reports accessible to authorized users.

3) The device as claimed in claim 1, wherein the display unit 102 is adapted to provide randomized quizzes using previously learned vocabulary and display correct meanings upon user’s failure to recall, to reinforce learning.

4) The device as claimed in claim 1, wherein the expandable ring arrangement 107 includes a plurality of curved rods 107a connected via motorized hinges 107b and integrated with medical-grade fabric for safety and comfort during use.

5) The device as claimed in claim 1, wherein the speech therapy tool 106 includes a vibrating unit 106d with intensity dynamically adjusted in real-time based on therapeutic requirements.

6) The device as claimed in claim 1, wherein the microcontroller is trained to analyze the child’s pronunciation accuracy over time and modify the vocabulary and exercise complexity accordingly to sustain engagement and improvement.

7) The device as claimed in claim 1, wherein the microphone unit 103 captures speech for NLP-based pronunciation comparison and generates immediate audio-visual feedback for speech correction.

8) The device as claimed in claim 1, wherein the cloud database stores weekly performance reports summarizing vocabulary learned, quiz accuracy, and areas requiring improvement, and transmits the report to authorized guardians or teachers.

9) The device as claimed in claim 1, wherein the microcontroller is adapted to dynamically adjust training difficulty based on user progress, analyze performance data, and issue personalized feedback.

10) The device as claimed in claim 1, wherein the microcontroller displays pitch modulation exercises and dynamic tongue twisters with gradually increasing difficulty levels to improve articulation and control under pressure.

Documents

Application Documents

# Name Date
1 202541077315-STATEMENT OF UNDERTAKING (FORM 3) [13-08-2025(online)].pdf 2025-08-13
2 202541077315-REQUEST FOR EXAMINATION (FORM-18) [13-08-2025(online)].pdf 2025-08-13
3 202541077315-REQUEST FOR EARLY PUBLICATION(FORM-9) [13-08-2025(online)].pdf 2025-08-13
4 202541077315-PROOF OF RIGHT [13-08-2025(online)].pdf 2025-08-13
5 202541077315-POWER OF AUTHORITY [13-08-2025(online)].pdf 2025-08-13
6 202541077315-FORM-9 [13-08-2025(online)].pdf 2025-08-13
7 202541077315-FORM FOR SMALL ENTITY(FORM-28) [13-08-2025(online)].pdf 2025-08-13
8 202541077315-FORM 18 [13-08-2025(online)].pdf 2025-08-13
9 202541077315-FORM 1 [13-08-2025(online)].pdf 2025-08-13
10 202541077315-FIGURE OF ABSTRACT [13-08-2025(online)].pdf 2025-08-13
11 202541077315-EVIDENCE FOR REGISTRATION UNDER SSI(FORM-28) [13-08-2025(online)].pdf 2025-08-13
12 202541077315-EVIDENCE FOR REGISTRATION UNDER SSI [13-08-2025(online)].pdf 2025-08-13
13 202541077315-EDUCATIONAL INSTITUTION(S) [13-08-2025(online)].pdf 2025-08-13
14 202541077315-DRAWINGS [13-08-2025(online)].pdf 2025-08-13
15 202541077315-DECLARATION OF INVENTORSHIP (FORM 5) [13-08-2025(online)].pdf 2025-08-13
16 202541077315-COMPLETE SPECIFICATION [13-08-2025(online)].pdf 2025-08-13