Abstract: The present disclosure is related to mirror adjustment system in an automobile, which discloses method and system for dynamically adjusting one or more mirrors in vehicle. Method includes receiving, by mirror adjustment system, real-time images of user driving vehicle, from an image capturing device associated with mirror adjustment system and determining an eye point of user with respect to image capturing device, using real-time images. Thereafter, the system determines hip point of user based on determined eye point of user, and then values of set of pre-defined parameters corresponding to mirrors, are determined based on determined eye point of user and hip point of the user. Finally, the system controls actuators associated with mirrors, based on determined values of set of pre-defined parameters, to dynamically adjust position of mirrors corresponding to field of view of user. FIG. 1
Claims:WE CLAIM:
1. A method for dynamically adjusting one or more mirrors (108) in a vehicle (240), the method comprising:
receiving, by a mirror adjustment system (104), one or more real-time images of a user (242) driving the vehicle (240), from an image capturing device (102) associated with the mirror adjustment system (104);
determining, by the mirror adjustment system (104), an eye point of the user (242) with respect to the image capturing device (102), using the one or more real-time images;
determining, by the mirror adjustment system (104), a hip point of the user (242) based on the determined eye point of the user (242);
determining, by the mirror adjustment system (104), values of a set of pre-defined parameters corresponding to one or more mirrors (108), based on the determined eye point of the user (242) and the hip point of the user (242); and
controlling, by the mirror adjustment system (104), actuators (106) associated with the one or more mirrors (108), based on the determined values of set of pre-defined parameters, to dynamically adjust position of the one or more mirrors (108) corresponding to a field of view of the user (242).
2. The method as claimed in claim 1 comprises determining association between each eye point and a pre-defined hip point, based on a first pre-defined technique.
3. The method as claimed in claim 1, wherein the set of pre-defined parameters comprises a mirror orientation, a vertical mirror adjustment angle, and a horizontal mirror adjustment angle.
4. The method as claimed in claim 1, wherein the values of the set of pre-defined parameters are determined for each pair of the eye point and the corresponding hip point based on a second pre-defined technique.
5. The method as claimed in claim 1 comprises:
receiving, by a gesture detection unit (116) associated with the mirror adjustment system (104), one or more images related to a gesture performed by the user (242), from the image capturing device (102);
detecting, by the gesture detection unit (116), identity of the user (242) and the gesture performed by the user (242) based on the one or more images;
retrieving, by the gesture detection unit (116), a user profile corresponding to the user (242) detected in the received images;
mapping, by the gesture detection unit (116), the gesture detected in the one or more images with one or more pre-defined gestures associated with the user profile, wherein each of the one or more pre-defined gestures is associated with a unique predefined action related to one or more systems configured in the vehicle (240);
determining, by the gesture detection unit (116), the pre-defined action corresponding to the gesture performed by the user (242), based on the mapping; and
performing, by the gesture detection unit (116), the determined pre-defined action related to the one or more systems, in response to the gesture performed by the user (242).
6. The method as claimed in claim 5, wherein the gesture comprises at least one of hand gestures, eye gestures and head-movement based gestures.
7. A mirror adjustment system (104) for dynamically adjusting one or more mirrors (108) in a vehicle (240), the mirror adjustment system (104) comprising:
a processor (110); and
a memory (114) communicatively coupled to the processor (110), wherein the memory (114) stores the processor instructions, which on execution, causes the processor (110) to:
receive one or more real-time images of a user (242) driving the vehicle (240), from an image capturing device (102) associated with the mirror adjustment system (104);
determine an eye point of the user (242) with respect to the image capturing device (102), using the one or more real-time images;
determine a hip point of the user (242) based on the determined eye point of the user (242);
determine values of a set of pre-defined parameters corresponding to one or more mirrors (108), based on the determined eye point of the user (242) and the hip point of the user (242); and
control actuators (106) associated with the one or more mirrors (108), based on the determined values of set of pre-defined parameters, to dynamically adjust position of the one or more mirrors (108) corresponding to a field of view of the user (242).
8. The mirror adjustment system (104) as claimed in claim 7, wherein, the processor (110) is configured to determine an association between each eye point and a pre-defined hip point, based on a first pre-defined technique.
9. The mirror adjustment system (104) as claimed in claim 7, wherein the set of pre-defined parameters comprises a mirror orientation, a vertical mirror adjustment angle, and a horizontal mirror adjustment angle.
10. The mirror adjustment system (104) as claimed in claim 7, wherein the values of the set of pre-defined parameters are determined for each pair of the eye point and the corresponding hip point based on a second pre-defined technique.
11. The mirror adjustment system (104) as claimed in claim 7 comprises a gesture detection unit (116), wherein the gesture detection unit (116) is configured to:
receive one or more images related to a gesture performed by the user (242), from the image capturing device (102);
detect identity of the user (242) and the gesture performed by the user (242) based on the one or more images;
retrieve a user profile corresponding to the user (242) detected in the received images;
map the gesture detected in the one or more images with one or more pre-defined gestures associated with the user profile, wherein each of the one or more pre-defined gestures is associated with a unique predefined action related to one or more systems configured in the vehicle (240);
determine the pre-defined action corresponding to the gesture performed by the user (242), based on the mapping; and
perform the determined pre-defined action related to the one or more systems, in response to the gesture performed by the user (242).
12. The gesture detection unit (116) as claimed in claim 11, wherein the gesture comprises at least one of hand gestures, eye gestures and head-movement based gestures.
, Description:TECHNICAL FIELD
[0001] The present subject matter relates generally to a mirror adjustment system in automobiles. More particularly, but not exclusively, discloses a method and a system for dynamically adjusting one or more mirrors in a vehicle.
BACKGROUND
[0002] Generally, automotive industry makes considerable effort to improve safety of vehicles and occupants of the vehicle. One of the areas of concern has been, for instance, a limited field-of-view provided by the rear-view mirrors including both Inside Rear View Mirror (IRVM) and Outside Rear View Mirror (ORVM) of the vehicle. Generally, rear-view mirrors of the vehicle are oriented to provide the driver of the vehicle with an optimum rear viewing zone behind and towards the left-side and towards the right-side of the vehicle. For manually adjustable rear-view mirrors, the rear viewing zone is a fixed region of space in the vicinity of the vehicle, which is at a particular orientation relative to the vehicle. The orientation of the rear viewing zone can be adjusted by adjusting the rear-view mirrors.
[0003] Effective use of vehicle mirrors is an essential part of safe driving. However, in some instances, a vehicle mirror may be disoriented which requires manual adjustment. In some other instances, the vehicle mirrors may become ineffective if the driver repositions his seating, for example, moves forward/backward by adjusting the seat under certain driving conditions, or if the driver of the vehicle itself changes, since dimension profile of each driver is different. This may change the driver's viewing perspective in the rear-view mirror and may cause the rear-view mirror to be ineffective, as the driver can no longer properly view vehicles behind the vehicle. Furthermore, manually reaching up to the rear-view mirrors to adjust them while driving may be dangerous, as it deviates the driver's attention from operation of the vehicle.
[0004] One existing system addresses the above-mentioned problem by automatically adjusting the mirror of vehicle, by identifying a three-dimensional position of a reference point such as eye of the driver. Since, the driver’s position and dimensions are not considered for mirror adjustment computation, using such systems to adjust the mirrors of the vehicle may lead to incorrect adjustment angles, which may not be apt for the driver. Such incorrect adjustment may not only provide incorrect view to the driver, but sometimes may even block the rear-view partially or completely, or may even divert the driver’s attention from driving, leading to accidents. Another existing system provides an automatic adjustment of mirrors based on the position of the driver’s seat. However, such systems as well may not appropriately adjust the mirrors of the vehicle since, the driver dimensions and eye point of the driver are not considered, which may vary for each driver.
[0005] Additionally, the driver in the vehicle may interact with physical controls such as knobs, dials, and switches on a console inside the vehicle, to control various vehicle features related to infotainment system, power windows, Heating-Ventilation-and-Air conditioning (HVAC) system and the like. Controlling various systems of the vehicle using physical controls may distract the driver while driving the vehicle. For instance, operating infotainment system, operating windows, operating mobile phone features connected to the infotainment system, adjusting mirrors manually and the like, may be highly distracting to the driver. One existing system allows the user to perform physical gestures on a touchscreen that displays a user interface. However, it is often cumbersome and inconvenient for the user to reach forward or sideways to interact with a touchscreen or manipulate a physical control, and these existing systems frequently present the user with a large number of functions that can be confusing and difficult to use. In another existing system, the driver controls a component of a vehicle by performing a gesture in front of a camera. However, the physical gestures on the touch screen and the gesture in front of camera for performing various functionalities of the vehicle may be pre-configured at the time of manufacture of the vehicle. This may be confusing and difficult to remember for the user and, also the driver may not be comfortable with certain pre-configured gestures. This in turn may distract the driver more under certain driving conditions or may cause the driver to not use the gestures at all, both of which are disadvantageous.
[0006] The information disclosed in this background of the disclosure section is only for enhancement of understanding of the general background of the invention and should not be taken as an acknowledgment or any form of suggestion that this information forms the prior art already known to a person skilled in the art.
SUMMARY
[0007] Disclosed herein is a method for dynamically adjusting one or more mirrors in a vehicle. The method includes receiving one or more real-time images of a user driving a vehicle, from an image capturing device associated with a mirror adjustment system. Further, the method includes determining an eye point of the user with respect to the image capturing device, using the one or more real-time images. Furthermore, the method includes determining a hip point of the user based on the determined eye point of the user. Thereafter, the method includes determining values of a set of pre-defined parameters corresponding to one or more mirrors, based on the determined eye point of the user and the hip point of the user. Finally, the method includes controlling actuators associated with the one or more mirrors, based on the determined values of set of pre-defined parameters, to dynamically adjust position of the one or more mirrors corresponding to a field of view of the user.
[0008] In an embodiment, the method further includes determining association between each eye point and a pre-defined hip point, based on a first pre-defined technique.
[0009] In an embodiment, the set of pre-defined parameters includes, at least one of, a mirror orientation, a vertical mirror adjustment angle, and a horizontal mirror adjustment angle.
[0010] In an embodiment, the values of the set of pre-defined parameters are determined for each pair of the eye point and the corresponding hip point based on a second pre-defined technique.
[0011] Disclosed herein is a mirror adjustment system for dynamically adjusting one or more mirrors in a vehicle. The mirror adjustment system comprises a processor and a memory communicatively coupled to the processor. The memory stores the processor instructions, which on execution, causes the processor to receive one or more real-time images of a user driving a vehicle, from an image capturing device associated with a mirror adjustment system. Further, the processor determines an eye point of the user with respect to the image capturing device, using the one or more real-time images. Furthermore, the processor determines a hip point of the user based on the determined eye point of the user. Thereafter, the processor determines values of a set of pre-defined parameters corresponding to one or more mirrors, based on the determined eye point of the user and the hip point of the user. Finally, the processor controls actuators associated with the one or more mirrors, based on the determined values of set of pre-defined parameters, to dynamically adjust position of the one or more mirrors corresponding to a field of view of the user.
[0012] In some implementations, the method further includes receiving one or more images related to a gesture performed by the user, from the image capturing device. Further, the method includes detecting identity of the user and the gesture performed by the user based on the one or more images. Furthermore, the method includes retrieving a user profile corresponding to the user detected in the received images. Thereafter, the method includes mapping the gesture detected in the one or more images with one or more pre-defined gestures associated with the user profile. Each of the one or more pre-defined gestures is associated with a unique predefined action related to one or more systems configured in the vehicle. Further, the method includes determining the pre-defined action corresponding to the gesture performed by the user, based on the mapping. Finally, the method includes performing the determined pre-defined action related to the one or more systems, in response to the gesture performed by the user.
[0013] In an embodiment, the gesture includes at least one of hand gestures, eye gestures and head-movement based gestures.
[0014] In some implementations, the mirror adjustment system is associated with a gesture detection unit. The gesture detection unit receives one or more images related to a gesture performed by the user, from the image capturing device. Further, the gesture detection unit detects identity of the user and the gesture performed by the user based on the one or more images. Furthermore, the gesture detection unit retrieves a user profile corresponding to the user detected in the received images. Thereafter, the gesture detection unit maps the gesture detected in the one or more images with one or more pre-defined gestures associated with the user profile. One or more pre-defined gestures is associated with a unique predefined action related to one or more systems configured in the vehicle. Further, the gesture detection unit determines the pre-defined action corresponding to the gesture performed by the user, based on the mapping. Finally, the gesture detection unit performs the determined pre-defined action related to the one or more systems, in response to the gesture performed by the user.
[0015] The foregoing summary is illustrative only and is not intended to be in any way limiting. In addition to the illustrative aspects, embodiments, and features described above, further aspects, embodiments, and features will become apparent by reference to the drawings and the following detailed description.
BRIEF DESCRIPTION OF THE ACCOMPANYING DIAGRAMS
[0016] The accompanying drawings, which are incorporated in and constitute a part of this disclosure, illustrate exemplary embodiments and, together with the description, serve to explain the disclosed principles. In the figures, the left-most digit(s) of a reference number identifies the figure in which the reference number first appears. The same numbers are used throughout the figures to reference like features and components. Some embodiments of system and/or methods in accordance with embodiments of the present subject matter are now described, by way of example only, and with reference to the accompanying figures, in which:
[0017] FIG. 1A illustrates an exemplary architecture for dynamically adjusting one or more mirrors in a vehicle, in accordance with some embodiments of the present disclosure;
[0018] FIG. 1B illustrates an exemplary architecture along with gesture detection unit for performing pre-defined actions related to the one or more systems, in response to the gestures performed by a user, in accordance with some embodiments of the present disclosure;
[0019] FIG. 2A illustrates a detailed block diagram of a mirror adjustment system for dynamically adjusting one or more mirrors in a vehicle, in accordance with some embodiments of the present disclosure;
[0020] FIG. 2B illustrates a schematic diagram of exemplary reference points to be determined within a vehicle to dynamically adjust one or more mirrors, in accordance with some embodiments of the present disclosure;
[0021] FIG. 2C illustrates a schematic diagram of exemplary vertical mirror adjustment scenario, in accordance with some embodiments of the present disclosure;
[0022] FIG. 2D illustrates a schematic diagram of exemplary horizontal mirror adjustment scenario, in accordance with some embodiments of the present disclosure;
[0023] FIG. 2E illustrates a schematic diagram of exemplary scenario of user viewing the inner rear-view mirror, upon dynamic adjustment, in accordance with some embodiments of the present disclosure;
[0024] FIG. 2F illustrates a schematic diagram of exemplary scenario where user is performing a gesture for receiving/rejecting a call, in accordance with some embodiments of the present disclosure;
[0025] FIG. 3A is a flowchart depicting a method for dynamically adjusting one or more mirrors in a vehicle, in accordance with some embodiments of the present disclosure;
[0026] FIG. 3B is a flowchart depicting a method for performing pre-defined actions related to the one or more systems, in response to the gestures performed by a user, in accordance with some embodiments of the present disclosure; and
[0027] FIG. 4 is a block diagram of an exemplary computer system for implementing embodiments consistent with the present disclosure.
[0028] It should be appreciated by those skilled in the art that any block diagrams herein represent conceptual views of illustrative systems embodying the principles of the present subject matter. Similarly, it will be appreciated that any flow charts, flow diagrams, state transition diagrams, pseudo code, and the like represent various processes which may be substantially represented in computer readable medium and executed by a computer or processor, whether or not such computer or processor is explicitly shown.
DETAILED DESCRIPTION
[0029] In the present document, the word “exemplary” is used herein to mean “serving as an example, instance, or illustration”. Any embodiment or implementation of the present subject matter described herein as “exemplary” is not necessarily to be construed as preferred or advantageous over other embodiments.
[0030] While the disclosure is susceptible to various modifications and alternative forms, specific embodiment thereof has been shown by way of example in the drawings and will be described in detail below. It should be understood, however that it is not intended to limit the disclosure to the forms disclosed, but on the contrary, the disclosure is to cover all modifications, equivalents, and alternative falling within the scope of the disclosure.
[0031] The terms “comprises”, “comprising”, “includes” or any other variations thereof, are intended to cover a non-exclusive inclusion, such that a setup, device, or method that includes a list of components or steps does not include only those components or steps but may include other components or steps not expressly listed or inherent to such setup or device or method. In other words, one or more elements in a system or apparatus proceeded by “comprises… a” does not, without more constraints, preclude the existence of other elements or additional elements in the system or method.
[0032] Disclosed herein is a method and mirror adjustment system for dynamically adjusting one or more mirrors in a vehicle. The mirror adjustment system may automatically change the viewing angle of rear-view mirrors of the vehicle for better viewing experience, in response to determination of an eye point and a hip point of a user driving the vehicle. It should be noted that by detecting the position of the user’s eye and in turn calculating the hip point, the mirror adjustment system can automatically set the mirrors to an optimal position which provides a proper rear view to the user (driver). Upon activating an image capturing device mounted in the vehicle for capturing images of the user, the mirror adjustment system may determine an eye point of the user using the captured images. The mirror adjustment system may subsequently determine user’s hip point based on the user’s eye level (eye point), with respect to interior of the vehicle, for instance, cabin of the driver seat. By correlating determined information related to the eye point and the hip point, the mirror adjustment system may determine parameters such as a vertical angle by which the mirror should be moved, horizontal angle by which the mirror should be moved, orientation of the mirror and the like, which may provide an optimal rear view apt for the driver , while minimizing the size of the “blind spot” that can occur. The “blind spot” may be a scenario when the objects or road in the rear view of the driver appear to be hidden/invisible in a field of view to the driver. Thereafter, the mirror adjustment system adjusts one or more mirrors to the determined orientation and angles with a minimal blind spot in accordance with the user's eye position and hip position.
[0033] Further, the mirror adjustment system may be communicatively coupled to a gesture detection unit. The gesture detection unit, upon receiving the images from the image capturing device, may map the gesture detected in the one or more images with one or more pre-defined gestures associated with a user profile. Each of the one or more pre-defined gestures may be associated with a unique predefined action related to one or more systems configured in the vehicle. As an example, the one or more systems may include, but not limited to, a window motor controlling system, a tailgate opening system, a fuel tank opening system, a Heating-Ventilation-Air Conditioning (HVAC) system, wiper activating system, and the like.
[0034] Based on the mapping, the gesture detection unit may determine the pre-defined action corresponding to the gesture performed by the user. In response to the gesture performed by the user, the gesture detection unit may perform the determined pre-defined action related to the one or more systems.
[0035] Embodiments herein enable automatic adjustment of one or more mirrors using eye point and hip point of the user driving the vehicle, which advantageously aids in providing appropriate field of view to the user without blind spot. By combining the eye point and hip point (height /position of the user), the exact distance from the eyes to the rear-view mirrors, an appropriate vertical and horizontal angle of the one or more mirrors can be determined, that will allow the user driving the vehicle to see proper rear view, while minimizing the size of the “blind spot” that can occur. Use of the mirror adjustment system disclosed in the embodiments herein aids in providing accuracy in positioning of the rear-view mirrors to reflect objects in the field of view of the user driving the vehicle for proper mirror settings are greatly increased. This results in eliminating manual adjustment of the mirror while driving or improper adjustments of the mirror, and in turn leading to lower frequency of accidents and injury. Since, the mirrors are appropriately adjusted to each user, there is more accuracy of mirror adjustment which seems appropriate to the field of view to the user and the seat position of the user. Further, the present disclosure provides a feature wherein the one or more mirrors of the vehicle are continuously adjusted in real-time or adjusted in real-time upon occurrence of events such as seat repositioning, change in driver and the like, which not only completely eliminates manual adjustment of the mirrors, but also provides the flexibility of auto adapting the mirrors to changing environment such as various drivers, various seat positioning and the like.
[0036] Embodiments herein allows the user to customize the gestures which is more suitable and easily remembered by the user to control the one or more systems in the vehicle. Thus, by performing the customized gestures to control the one or more systems, the user can concentrate on driving the vehicle more efficiently, in turn leading to lesser distraction, and lower frequency of accidents and injury.
[0037] A description of an embodiment with several components in communication with each other does not imply that all such components are required. On the contrary, a variety of optional components are described to illustrate the wide variety of possible embodiments of the invention.
[0038] In the following detailed description of the embodiments of the disclosure, reference is made to the accompanying drawings that form a part hereof, and in which are shown by way of illustration specific embodiments in which the disclosure may be practiced. These embodiments are described in sufficient detail to enable those skilled in the art to practice the disclosure, and it is to be understood that other embodiments may be utilized and that changes may be made without departing from the scope of the present disclosure. The following description is, therefore, not to be taken in a limiting sense.
[0039] FIG. 1A illustrates an exemplary architecture for dynamically adjusting one or more mirrors in a vehicle, in accordance with some embodiments of the present disclosure.
[0040] The architecture 100 includes an image capturing device 102, a mirror adjustment system 104, an actuator 1061 to an actuator 106n (collectively herein after referred as actuators 106), a mirror 1081 to a mirror 108n (collectively herein after referred as one or more mirrors 108). In some embodiments, the mirror adjustment system 104 may be an Electronic Control Unit (ECU) configured in the vehicle and associated with a Body Control Unit (BCU) (not shown in FIG. 1A) configured in the vehicle via a communication network. For instance, the vehicle may include at least one of, but not limited to, 2-wheelers, 3-wheelers, 4-wheelers, 6-wheelers, 8 wheelers. For example a motor bike, a car, a bus, a truck, a van, crossover vehicle, multi-purpose vehicle, sports utility vehicle, adventure utility vehicle, commercial vehicle including multi-axle vehicle, and the like. In some embodiments, the mirror adjustment system 104 may be associated with a Telematics/Transmission Control Unit (TCU), a Door Control Unit (DCU), a Human machine interface (HMI), a Powertrain Control Module (PCM), a Speed Control Unit (SCU), a Battery Management System (BMS), a Driver Monitoring System (DMS), an infotainment system, and the like (not shown in the FIG. 1A). In some implementations, the mirror adjustment system 104 could be implemented using an electronic device such as a mobile, a smart phone, a tablet, phablet, a personal digital assistant, and the like. When implemented mirror adjustment system 104 as an electronic device, then the electronic device may control the adjustment of position of the one or more mirrors in the vehicle through one of the devices such as Electronic Control Unit (ECU), Telematics/Transmission Control Unit (TCU), a Human machine interface (HMI), an infotainment system and the like by establishing suitable communicate interface known in the art.
[0041] Further, the image capturing device 102 may be communicatively coupled to the mirror adjustment system 104 via the communication network (not shown in FIG. 1A). The mirror adjustment system 104 may be communicatively coupled to the actuators 106 via the communication network (not shown in FIG. 1A). In some embodiments, the communication network may be at least one of a wired communication network and a wireless communication network. For illustrative purpose, the architecture 100 includes one image capturing device 102, however, it is to be understood that the architecture 100 may include, but not limited to, more than one image capturing device 102. As an example, the image capturing device 102 may include at least one of, but not limited to, a camera, a depth camera, an infrared camera, a smart camera, a 360-degree camera, a night vision camera, Red Green Blue (RGB) camera, three-dimensional (3D) camera, an eye recognition camera, a sensor, and the like. The image capturing device 102 may be aligned suitably inside the vehicle to capture an image of the user periodically. For instance, the image capturing device 102 may be placed within the user compartment (for instance driver) of the vehicle that allows the image capturing device 102 to capture the face of the driver without obstruction, e.g., above the steering, on the windshield, dashboard, headliner, or the like. The term user and driver may be interchangeably used in the foregoing description for better understanding of the description and exemplary scenarios. The image capturing device 102 may be effective to detect eyes of the user even when, for instance, the user is wearing glasses, sunglasses, or when the user is driving in the dark. As an example, the image capturing device 102 may capture image and/or video of a user (for instance driver) in front of the image capturing device 102 or the interior of the vehicle. In some implementations, the image capturing device 102 may automatically detect change in head/body position of the user, for instance, sideways, forward, backwards and the like. For example, the head/body position may be detected to determine initial mirror adjustment setting, such as, for example, when the user enters the vehicle, when starting the vehicle, and the like. The initial mirror adjustment setting may be used as the appropriate rear-view setting. Changes in the head/body position of the user may be detected in predetermined temporal increments, such as, continuously, or once every 15 minutes or less, and so on, or upon the occurrence of an event such as, for example, adjustment of the driver seat, change in the driver and the like. Upon determining the change in head/body position of the user, the mirror adjustment system 104 may automatically adjust the one or more mirrors 108 to maintain an appropriate field of view of the user.
[0042] Further, the actuators 106 may be coupled to one or more motors (not shown in FIG. 1A) for adjusting the one or more mirrors 108. In an instance, each mirror assembly may include, at least one of, but not limited to, a motor and a reflective element such as a mirror operatively coupled with the motor. The motors may be used to adjust the angle of the one or more mirrors 108 and thereby the view provided by the one or more mirrors 108. The actuators 106 may be configured within one or more mirror assemblies of the one or more mirrors 108 or may configured as stand-alone unit in the vehicle. For instance, the one or more mirrors 108 may be at least one of, but not limited to, an Inner Rear-View Mirror (IRVM), an Outer Rear-View Mirror (ORVM), and the like. The ORVM may include at least one of, but not limited to, a right-side rear-view mirror and a left side rear-view mirror.
[0043] Further, the mirror adjustment system 104 may include a processor 110, an Input/Output (I/O) interface 112 and a memory 114. The I/O interface 112 may be configured to receive image data comprising one or more real-time images and videos, from the image capturing device 102 associated with the mirror adjustment system 104.
[0044] In some embodiments, the processor 110 may determine an eye point of the user with respect to the image capturing device 102, using the one or more real-time images. In some embodiments, the processor 110 may receive depth information related to both the eyes of the user from the image capturing device 102. In other words, the processor 110 may receive distance between each eye of the user and the image capturing device 102, from the image capturing device 102. Thereafter, the processor 110 may determine the eye point of the user by computing an average distance of both the eyes of the user from the image capturing device 102. In some embodiments, the eye point of the user may be in the form of (X, Y, Z) co-ordinates, wherein “X” co-ordinate indicates longitudinal dimension, “Y” co-ordinate indicates lateral dimension, and “Z” co-ordinate indicates vertical dimension. Furthermore, the processor 110 may determine a hip point of the user based on the determined eye point of the user. The hip point of the user is the point where the hip portion/joint of the user is resting on the seat. In an embodiment, the processor 110 may determine the hip point based on an association between each eye point and a pre-defined hip point, determined using a first pre-defined technique.
[0045] Thereafter, the processor 110 may determine values of a set of pre-defined parameters corresponding to one or more mirrors 108, based on the determined eye point of the user and the hip point of the user. In some embodiments, the set of pre-defined parameters may include at least one of, but not limited to, a mirror orientation, a vertical mirror adjustment angle, and a horizontal mirror adjustment angle, and the like.
[0046] In some embodiments, the values of the set of pre-defined parameters are determined for each pair of the eye point and the corresponding hip point based on a second pre-defined technique. Thereafter, the processor 110 may control actuators 106 associated with the one or more mirrors 108, based on the determined values of set of pre-defined parameters, to dynamically adjust position of the one or more mirrors 108 corresponding to a field of view of the user. For instance, depending on the determined adjustment angles for the left side rear-view mirror, using the processor 110, a left side motor may adjust the left mirror receiving signals from the actuators 106. Likewise, depending on the determined adjustment angles for the right-side rear-view mirror, using the processor 110, a right-side motor may adjust the right mirror receiving signals from the actuators 106.
[0047] In some embodiments, the mirror adjustment system 104 may be associated with a gesture detection unit 116 as shown in architecture 101 in FIG. 1B. In some embodiments, the gesture detection unit 116 may be configured in the vehicle. In some other embodiments, the gesture detection system 116 may be an external system. Further, the gesture detection unit 116, may include a processor, an Input/Output (I/O) interface and a memory (not shown in FIG. 1B).
[0048] In some embodiments, the gesture detection unit 116 may receive one or more images related to a gesture performed by the user, from the image capturing device 102. In some embodiments, the gesture detection unit 116 may receive the one or more images from the mirror adjustment system 104, which in turn receives the one or more images from the image capturing device 102. The gesture detection unit 116 may detect identity of the user and the gesture performed by the user based on the one or more images. In some embodiments, the gesture may include, at least one of, but not limited to, hand gestures, eye gestures and head-movement based gestures. In some embodiments, the gesture detection unit 116 may detect the gesture using one or more predefined gesture detecting techniques. As an example, the one or more predefined gesture detecting techniques may include, but not limited to, Artificial Intelligence (AI) based techniques such as machine learning techniques, deep learning techniques and the like, and non-AI based techniques. Further, the gesture detection unit 116 may retrieve a user profile corresponding to the user detected in the received images. In some embodiments, the user profiles may be stored in the memory 114 or may be stored in a database associated with the gesture detection unit 116. Furthermore, the gesture detection unit 116 may map the gesture detected in the one or more images with one or more pre-defined gestures associated with the user profile. In some embodiments, each of the one or more pre-defined gestures may be associated with a unique predefined action related to one or more systems configured in the vehicle. The one or more systems may include at least one of, but not limited to, a window motor controlling system, a tailgate opening system, a fuel tank opening system, a wiper activating system, and the like.
[0049] Further, the gesture detection unit 116 may determine the pre-defined action corresponding to the gesture performed by the user, by mapping the detected gesture with customized predefined gestures and predefined actions. Thereafter, the gesture detection unit 116 may perform the determined pre-defined action related to the one or more systems, in response to the gesture performed by the user.
[0050] FIG. 2A illustrates a detailed block diagram of a mirror adjustment system 104 for dynamically adjusting one or more mirrors 108 in a vehicle, in accordance with some embodiments of the present disclosure.
[0051] In some implementations, the mirror adjustment system 104 may include data 202 and modules 204. As an example, the data 202 is stored in the memory 114 configured in mirror adjustment system 104 as shown in the FIG. 2A. In one embodiment, the data 202 may include image data 206, eye point data 208, hip point data 210, parametric data 212, and other data 214.
[0052] In an embodiment, the data 202 may be stored in the memory 114 in the form of various data structures. Additionally, the data 202 can be organized using data models, such as relational or hierarchical data models. The other data 214 may store data, including temporary data and temporary files, generated by the modules 204 for performing the various functions of the mirror adjustment system 104.
[0053] In an embodiment, the data 202 stored in the memory 114 may be processed by the modules 204 of the mirror adjustment system 104. The modules 204 may be stored within the memory 114. In an example, the modules 204 communicatively coupled to the processor 110 configured in the mirror adjustment system 104, may also be present outside the memory 114 as shown in FIG. 2A and implemented as hardware. As used herein, the term modules refer to an application-specific integrated circuit (ASIC), an electronic circuit, a processor (shared, dedicated, or group) and memory that execute one or more software or firmware programs, a combinational logic circuit, and/or other suitable components that provide the described functionality.
[0054] In an embodiment, the modules 204 may include, for example, a receiving module 222, an eye point determining module 224, a hip point determining module 226, a parameter determining module 228, a controlling module 230, and other modules 232. The other modules 232 may be used to perform various miscellaneous functionalities of the mirror adjustment system 104. It will be appreciated that such aforementioned modules 204 may be represented as a single module or a combination of different modules.
[0055] In some embodiments, the receiving module 222 may receive one or more real-time images of a user driving a vehicle, from the image capturing device 102 associated with the mirror adjustment system 104. The one or more real-time images may be stored as image data 206. The one or more real-time images of the user may include eye/eyes or entire face portion or even body portion of the user driving the vehicle 240 as per the configuration/installation of the image capturing device 102.
[0056] Using the one or more real-time images, the eye point determining module 224 may determine an eye point of the user with respect to the image capturing device 102. The eye point of the user is an average distance of both the eyes from the image capturing device 102. The eye point of the user may be stored as eye point data 208. In one example, a location of the image capturing device 102 may be an origin (0,0,0) of coordinate system that depicts coordinates of the eye point. The coordinates may include, but not limited to, three coordinates (e.g., X, Y and Z). The coordinates (X, Y, Z) may identify a depth information of user’s both eye points i.e., depth of each eye of the user from the image capturing device 102, in X, Y, Z coordinates. The coordinate ‘X’ may be a longitudinal dimension, the coordinate ‘Y’ may be a lateral dimension and the coordinate ‘Z’ may be vertical dimension of the eye point of the user. The average of the distance of two eyes of the user from the image capturing device 102 may be determined as an eye point of the user. For example, the eye point of the user from the image capturing device 102, may be for instance, Eye point in X dimension may be E (X)=+500 milli-meter (mm), Eye point in Y dimension may be E (Y) = 40 mm and the Eye point in Z dimension may be E (Z)= 44 mm. A positive value for “X” dimension may indicate a forward direction from a reference point, for “Y” dimension may indicate an outboard direction (i.e. direction right/left) from the reference point (i.e., to the user’s side), and for “Z” dimension may indicate direction above the reference point. For example, coordinates (X, Y, Z) may identify a point 500 mm units away from image capturing device 102 in the positive X-direction, 40 mm units away from the image capturing device 102 in the positive Y-direction, and 44 mm units away from the image capturing device 102 in the positive Z-direction. In another example, coordinates may include one pair of coordinates representing a point between a left eye and a right eye of the user in the one or more real-time images.
[0057] Based on the determined eye point of the user, the hip point determining module 226 may determine a hip point of the user. The hip point of the user is the point where the hip portion/joint of the user is resting on the seat. The hip point of the user may be stored as hip point data 210. The image capturing device 102 preferably captures the eye portion or face portion of the user driving the vehicle 240. Further, the hip point determining module 226, preferably may determine a hip point of the user driving the vehicle 240 based on a predefined association between each eye point and a pre-defined hip point. As an example, the predefined associate may be in the form of a lookup table. Such association may be established based on a first pre-defined technique. In some embodiments, the first pre-defined technique may be a standard empirical equation that establishes an association between the eyepoint and the hip point of the user. Exemplary look-up table with exemplary values determined using a first predefined technique may be as shown in the below Table 1.
User’s eye point User’s hip point Step angle Right ORVM position Left ORVM position IRVM position
(500, 40, 44) -591 2° (6, 8) (5, -3) (-2,4)
(600, 43, 49) -586 2° (7, 5) (6, -2) (-3,7)
Table 1
[0058] The Table 1 includes calculated values for eye point of the user, hip point of the user, step angle, right ORVM positions, left ORVM positions, IRVM positions. The lookup table in Table 1 may represent the structure of data and mirror adjustment rotation angle information given to the actuators 106. In some embodiments, the look-up table may include, but not limited to, data that maps coordinates of eye point and hip point of the user with required mirror positions of the left ORVM, the right ORVM and the IRVM. For instance, consider that the image capturing device 102 rotates at step angle of 2° which may be specific to mirror design of the one or more mirrors 108 in both vertical and horizontal direction. The Table 1 contains all possible mirror positions of the one or more mirrors 108 in (a, b) format. In an example, default mirror location is (0, 0) which implies that the horizontal angle is ‘0’ and vertical angle is ‘0’. The current mirror positions may be stored in the memory 114 of the mirror adjustment system 104. For instance, the current mirror positions may be identified as (5, 8). This implies that current mirror position is moving 5 steps forward and rotating 8 steps upward from the default position. In other words, the moving 5 steps forward may be considered as 10° horizontal rotation angle and the rotating 8 steps upward rotation may be considered as 16° vertical rotation angle. The positions can also be negative indicating the direction change or rotation in an opposite direction.
[0059] As the driver always moves the eyes while driving only eye point input may make the ORVM rotation undesirable. Therefore, once the user sets the seat position then hip point may be constant. Considering the hip point as an input for adjusting the one or more mirrors 108 may provide a robust mirror adjustment angle and system performance. For instance, consider the hip point height of the user as H (Z). The H (Z) can be, for instance, E(Z) - 635 mm, if the percentage of face covered in pixels row of the image capturing device 102 is greater than 50 percentage height. Further, the H (Z) can be, for instance, E(Z) - 594 mm, if the percentage of face covered in pixels row of the image capturing device 102 is less than 50 percentage height. This subtraction factor may change based on the design of the vehicle 240. In some embodiments, the values 635mm and 594mm may be values with reference to Society of Automotive Engineers (SAE) standard, for instance, SAEJ286. In some embodiments, these values may also be configured based on vehicle dimension during development stage.
[0060] In some embodiments, the parameter determining module 228 may determine values of a set of pre-defined parameters corresponding to one or more mirrors 108, based on the determined eye point of the user and the hip point of the user. In some embodiments, the set of pre-defined parameters may include, but not limited to, a mirror orientation, a vertical mirror adjustment angle, and a horizontal mirror adjustment angle, and the like, and stored as parametric data 212. In some embodiments, the values of the set of pre-defined parameters are determined for each pair of the eye point and the corresponding hip point based on a second pre-defined technique. In some embodiments, the vertical mirror adjustment may be performed by adjusting a vertical angle (F) as shown in the FIG. 2C. In some embodiments, the vertical angle may be an angle formed between a reference line and a line along mirror aiming direction as shown in the FIG.2C. In some embodiments, the reference line of the vehicle 240 is constant as shown in FIG. 2C and 2D. As it is well known, the user 242 driving the vehicle 240 intuitively looks at the centre of the one or more mirrors 108. The intersection of the mirror's X, Y, Z axes/dimension constitutes a reference line, which remains stationary with respect to the housing regardless of mirror adjustment. Hereafter, the location of the mirror reference line may be considered coincident with the centre of the mirror and the mirror pivot point.
[0061] In some embodiments, the nominal location (referenced to the centre of the mirror) of the field of view of the user 242 may be a function of at least one of, but not limited to, the width (end to end) and height (top to bottom) of the ORVM, known vehicle dimensions, location where the ORVM housing is mounted to the vehicle 240, angular field of view of the IRVM, and ranges of possible seat heights, seat back angle adjustments, seat positions, and a user height. With consideration of these variables, the parameter determining module 228 may determine a best estimate for a rotation direction of the one or more mirrors 108from the reference line to the nominal centre of the field of view of the user 242.
[0062] The mirror angle for the user in some instances, needs vertical adjustment of the mirror. Therefore, based on the determined eye point and hip point of the user, the mirror needs to be adjusted vertically. Similarly, based on the determined eye point and hip point of the user, the mirror needs to be adjusted horizontally. In some embodiments, the horizontal mirror adjustment by adjusting a horizontal angle (?) is as shown in the FIG. 2D. In some embodiments, the horizontal angle may be an angle formed between a reference line and a line along mirror aiming direction as shown in the FIG.2D. The reference line of the vehicle 240 may be constant as shown in FIG. 2C. The mirror angle for the user needs horizontal adjustment of the mirror. Considering the values of Table 1, if the eye point of the user is determined to be (500, 40, 44) in (X, Y, Z) dimensions, and the step angle of the image capturing device 102 is 2°, then the corresponding hip point may be (-591) as shown in the above Table 1. Therefore, based on the eye point and hip point, the positions of the right ORVM should be adjusted to move 6 steps forward and rotate 8 steps upward (i.e., 6, 8), the left ORVM should be adjusted to move 5 steps forward and rotate 3 steps downward (i.e., 5, -3), and the IRVM should be adjusted to move 2 steps downward and 4 steps forward direction (i.e., -2, 4). Similarly, in another example, if the eye point of the user is determined to be (600, 43, 49) in (X, Y, Z) dimensions, and the step angle of the image capturing device 102 is 2°, then the corresponding hip point could be (-586). Such hip point may be determined based on the first predefined technique, using the eye point of the user, and stored in the look-up table as shown in the above Table 1. Therefore, based on the eye point and hip point, the positions of the right LRVM should be adjusted to 7 steps forward and rotated 5 steps upward (i.e., 7, 5), the left ORVM should be adjusted to (6, -2), and the IRVM should be adjusted to (-3, 7). In some embodiments, the mirror angles such as the vertical angle and the horizontal angle may be determined by converting coordinates of the right ORVM position, the left ORVM position and the IRVM position provided in the look-up table such as Table 1, to the mirror angles. In some embodiments, such association between the co-ordinates and the mirror angles may be prepared by performing multiple trials using the vehicle with different users with the help of digital simulation. In some embodiments, the multiple trials may be conducted using vehicles and users in the real-world instead of digital simulation. The results thus obtained by performing multiple trials may be populated in the look-up table for a specific vehicle design. Similarly, look-up tables may be prepared for each type of vehicle design.
[0063] In some embodiments, the controlling module 230 may control actuators 106 associated with the one or more mirrors 108, based on the determined values of the set of pre-defined parameters, to dynamically adjust position of the one or more mirrors 108 corresponding to a field of view of the user. The process of determining whether to adjust one or more mirrors 108 is in some cases performed substantially continuously, e.g., on an ongoing basis, at least while the vehicle 240 is being operated. For instance, the user 242, may enter the vehicle 240 or might be driving the vehicle 240, the image capturing device 102, may be capturing the images of the user 242 continuously, as shown in FIG. 2E. In some other embodiments, the mirror adjustment system 104 may monitor a plurality of variables indicating change in head position of the user 242 or seat position of the user 242, to detect occurrence of events such as shift in position of the user 242 or seat of the user 242. Such events may change the field of view of the user 242 which in turn demands change in position of the one or more mirrors 108. Therefore, the mirror adjustment system 104 may dynamically determine values associated with the set of predefined parameters and dynamically adjust the one or more mirrors 108 automatically, upon detecting occurrence of such events. Upon dynamically adjusting position of the one or more mirrors 108 corresponding to a field of view of the user 242, the user may use the mirrors to get rear view of the vehicle 240 as shown in the FIG. 2E. In some embodiments,
[0064] In some embodiments, the mirror adjustment system 104 may be associated with the gesture detection unit 116. In some implementations, the gesture detection unit 116 may include data and modules (not shown in FIG. 2A). As an example, the data 205 is stored in the memory (not shown in FIG. 2A) configured in the gesture detection unit 116. In one embodiment, the data may include gesture data 216, user profile data 218, action data 220, and other data 221.
[0065] In an embodiment, the data may be stored in the memory of the gesture detection unit 116 in the form of various data structures. Additionally, the data can be organized using data models, such as relational or hierarchical data models. The other data 221 may store data, including temporary data and temporary files, generated by the modules for performing the various functions of the gesture detection unit 116 associated with the mirror adjustment system 104.
[0066] In an embodiment, the data stored in the memory may be processed by the modules of the gesture detection unit 116. The modules may be stored within the memory. In an example, the modules communicatively coupled to the processor (not shown in the FIG. 2A) configured in the gesture detection unit 116, may also be present outside the memory and implemented as hardware. As used herein, the term modules refer to an application-specific integrated circuit (ASIC), an electronic circuit, a processor (shared, dedicated, or group) and memory that execute one or more software or firmware programs, a combinational logic circuit, and/or other suitable components that provide the described functionality.
[0067] In some embodiments, the gesture detection unit 116 may receive one or more images related to a gesture performed by the user 242, from the image capturing device 102. In some embodiments, the gesture may include at least one of, but not limited to, hand gestures, eye gestures and head-movement based gestures, which may be stored as the gesture data 216. In some embodiments, the gesture includes three-dimensional coordinates of the extremities and joints in hand and forearm of the user 242. For example, the gesture data 216 for hand gestures may include coordinates representing the three-dimensional positions of user's elbow, wrist, and the fingertip and knuckles of each of the user's finger. Similarly, the gesture data 216 for eye gestures may include coordinates representing the three-dimensional positions of eyes, eye lids/lashes, eyeballs and the like. Further, the gesture data 216 for head-movement based gestures may include coordinates representing the three-dimensional positions of head, forehead and the like.
[0068] In some embodiments, the gesture detection unit 116 may detect identity of the user and the gesture performed by the user 242, based on the one or more images. In some embodiments, the gestures may be customized by the user 242. In some embodiments, the gesture detection unit 116 may retrieve a user profile corresponding to the user 242 detected in the received images. The user profile may include, but not limited to, user identity data, customized gestures, demographic data, likes/dislikes of the user, preferred/personalized settings, and the like. The user profile may be stored as the user profile data 218.
[0069] In some embodiments, the gesture detection unit 116 may map the gesture detected in the one or more images with one or more pre-defined gestures associated with the user profile. In some embodiment, each of the one or more pre-defined gestures is associated with a unique predefined action related to one or more systems configured in the vehicle 240. The one or more actions may include at least one of but not limited to, accepting/rejecting a call, increasing/decreasing volume, fuel level, heating/ventilation/air conditioning, start wiper, close windows, open fuel tank lid, and the like. The one or more actions may be stored as the action data 220. As an example, the one or more systems may include, but not limited to, a window motor controlling system, a tailgate opening system, a fuel tank opening system, a wiper activating system, a Heating-Ventilation-Air conditioning (HVAC) system, and the like. The below Table 2 shows exemplary predefined gestures and corresponding predefined actions.
Predefined gesture Predefined action
Swiping gesture from right to left Disconnect an incoming call
Swiping gesture from left to right Receive an incoming call
Moving hand in bottom up direction Increase volume
Moving hand in top down direction Decrease volume
Looping with an index finger in forward direction Increase temperature of AC
Looping with an index finger in backward direction Decrease temperature of AC
Table 2
[0070] In some embodiments, the gesture detection unit 116 may determine the pre-defined action corresponding to the gesture performed by the user 242, based on the mapping. In some embodiments, the gesture detection unit 116 may perform the determined pre-defined action related to the one or more systems, in response to the gesture performed by the user 242. For instance, the user 242 may perform a gesture at the imaging capturing device 102, to accept or reject an incoming call as shown in the FIG. 2F. In another example, the gesture may correspond to pre-defined actions to rotate the IRVM/ORVM to a particular orientation.
[0071] FIG. 3A is a flowchart depicting a method 300a for dynamically adjusting one or more mirrors 108 in a vehicle 240, in accordance with some embodiments of the present disclosure.
[0072] As illustrated in FIG. 3A, the method 300a includes one or more blocks illustrating a method of dynamically adjusting one or more mirrors 108 in the vehicle 240. The method 300a may be described in the general context of computer-executable instructions. Generally, computer-executable instructions can include routines, programs, objects, components, data structures, procedures, modules, and functions, which perform functions or implement abstract data types.
[0073] The order in which the method 300a is described is not intended to be construed as a limitation, and any number of the described method blocks can be combined in any order to implement the method 300a. Additionally, individual blocks may be deleted from the methods without departing from the spirit and scope of the subject matter described herein. Furthermore, the method 300a can be implemented in any suitable hardware, software, firmware, or combination thereof.
[0074] At block 302, the method 300a may include receiving, by the processor 110, one or more real-time images of a user driving a vehicle 240, from an image capturing device 102 associated with a mirror adjustment system 104.
[0075] At block 304, the method 300a may include determining, by the processor 110, an eye point of the user 242 with respect to the image capturing device 102, using the one or more real-time images. In some embodiments, eye portion or entire image of the face or the body portion of the user 242 may be captured by the image capturing device 102, based on configuration/installation of the image capturing device 102.
[0076] At block 306, the method 300a may include determining, by the processor 110, a hip point of the user 242 based on the determined eye point of the user 242. In some embodiments, the processor 110 may determine an association between each eye point and a pre-defined hip point, based on a first pre-defined technique. The hip point of the user 242 is the point where the hip portion of the user 242 is resting on the seat.
[0077] At block 308, the method 300a may include determining, by the processor 110, values of a set of pre-defined parameters corresponding to one or more mirrors 108, based on the determined eye point of the user 242 and the hip point of the user 242. In some embodiments, the set of pre-defined parameters may include, but not limited to, a mirror orientation, a vertical mirror adjustment angle, and a horizontal mirror adjustment angle. In some embodiments, the values of the set of pre-defined parameters may be determined for each pair of the eye point and the corresponding hip point based on a second pre-defined technique.
[0078] At block 310, the method 300a may include controlling, by the processor 110, actuators 106 associated with the one or more mirrors 108, based on the determined values of set of pre-defined parameters, to dynamically adjust position of the one or more mirrors 108 corresponding to a field of view of the user 242. In some embodiments, dynamically adjusting position of the one or more mirrors 108 is based on occurrence of events which may include, but not limited to, user 242 entering the vehicle 240, user 242 adjusting the seat of the vehicle 240, user 242 adjusting his/her position in the seat and the like. Additionally, the process of determining whether to adjust one or more mirrors 108 is in some cases may be performed substantially continuously, e.g., on an ongoing basis, at least while the vehicle 240 is being operated. For instance, the user 242, may enter the vehicle 240 or might be driving the vehicle 240, the image capturing device 102 may be capturing the images of the user 242 to check if dynamic adjustment of the one or more mirrors 108 is required.
[0079] FIG. 3B is a flowchart depicting a method 300b for performing pre-defined action related to the one or more systems, in response to the gesture performed by user 242, in accordance with some embodiments of the present disclosure.
[0080] As illustrated in FIG. 3B, the method 300b includes one or more blocks illustrating a method 300b for performing pre-defined action related to the one or more systems, in response to the gesture performed by user 242. The mirror adjustment system 104 is associated with a gesture detection unit 116. The method 300b may be described in the general context of computer-executable instructions. Generally, computer-executable instructions can include routines, programs, objects, components, data structures, procedures, modules, and functions, which perform functions or implement abstract data types.
[0081] The order in which the method 300b is described is not intended to be construed as a limitation, and any number of the described method blocks can be combined in any order to implement the method 300a. Additionally, individual blocks may be deleted from the methods without departing from the spirit and scope of the subject matter described herein. Furthermore, the method 300a can be implemented in any suitable hardware, software, firmware, or combination thereof.
[0082] At block 312, the method 300b may include receiving, by the gesture detection unit 116, one or more images related to a gesture performed by the user 242, from the image capturing device 102. In some embodiments, the gesture may include at least one of, but not limited to, hand gestures, eye gestures and head-movement based gestures. In some embodiments, the gesture includes three-dimensional coordinates of the extremities and joints in hand and forearm of the user 242. For example, the gesture data 216 for hand gestures may include coordinates representing the three-dimensional positions of user's elbow, wrist, and the fingertip and knuckles of each of the user's finger. In some embodiments, the gesture may be customized by the user 242. Similarly, the gesture data 216 for eye gestures may include coordinates representing the three-dimensional positions of eyes, eye lids/lashes, eyeballs, and the like. Further, the gesture data 216 for head-movement based gestures may include coordinates representing the three-dimensional positions of head, forehead, and the like.
[0083] At block 314, the method 300b may include detecting, by the gesture detection unit 116, identity of the user 242 and the gesture performed by the user 242 based on the one or more images.
[0084] At block 316, the method 300b may include retrieving, by the gesture detection unit 116, a user profile corresponding to the user 242 detected in the received images. The user profile may include, but not limited to, user identity data, customized gestures, demographic data, likes/dislikes of the user, preferred/personalized settings, and the like.
[0085] At block 318, the method 300b may include mapping, by the gesture detection unit 116, the gesture detected in the one or more images with one or more pre-defined gestures associated with the user profile. In some embodiments, each of the one or more pre-defined gestures is associated with a unique predefined action related to one or more systems configured in the vehicle 240. The one or more actions may include at least one of but not limited to, accepting/rejecting a call, increasing/decreasing volume, fuel level, heat/ventilation/air condition, start wiper, close windows, open fuel tank lid, and the like. In some embodiments, the one or more systems may include, but not limited to, window motor controlling system, tailgate opening system, fuel tank opening system, wiper activating system, Heating-Ventilation-Air condition (HVAC) system, and the like.
[0086] At block 320, the method 300b may include determining, by the gesture detection unit 116, the pre-defined action corresponding to the gesture performed by the user 242, based on the mapping.
[0087] At block 322, the method 300b may include performing, by the gesture detection unit 116, the determined pre-defined action related to the one or more systems, in response to the gesture performed by the user 242.
[0088] FIG. 4 is a block diagram of an exemplary computer system 400 for implementing embodiments consistent with the present disclosure.
[0089] In some embodiments, FIG. 4 illustrates a block diagram of an exemplary computer system 400 for implementing embodiments consistent with the present invention. In some embodiments, the computer system 400 can be a mirror adjustment system 104 that is used for dynamically adjusting one or more mirrors 108 in a vehicle 240, as shown in the FIG. 4. In some other embodiments, the computer system 600 can be a gesture detection unit 116 associated with the mirror adjustment system 104 to perform pre-defined action related to the one or more systems, in response to the gesture performed by user 242. The computer system 400 may include a central processing unit (“CPU” or “processor”) 402. The processor 402 may include at least one data processor for executing program components for executing user or system-generated business processes. A user may include a person, a person using a device such as those included in this invention, or such a device itself. The processor 402 may include specialized processing units such as integrated system (bus) controllers, memory management control units, floating-point units, graphics processing units, digital signal processing units, etc. The processor 402 may be disposed in communication with input devices 411 and output devices 412 via I/O interface 401. The I/O interface 401 may employ communication protocols/methods such as, without limitation, audio, analog, digital, stereo, IEEE-1395, serial bus, Universal Serial Bus (USB), infrared, PS/2, BNC, coaxial, component, composite, Digital Visual Interface (DVI), high-definition multimedia interface (HDMI), Radio Frequency (RF) antennas, S-Video, Video Graphics Array (VGA), IEEE 802.n /b/g/n/x, Bluetooth, cellular (e.g., Code-Division Multiple Access (CDMA), High-Speed Packet Access (HSPA+), Global System For Mobile Communications (GSM), Long-Term Evolution (LTE), WiMax, or the like), etc.
[0090] Using the I/O interface 401, computer system 400 may communicate with input devices 411 and output devices 412. In some embodiments, the processor 402 may be disposed in communication with a communication network 409 via a network interface 403. The network interface 403 may communicate with the communication network 409. The network interface 403 may employ connection protocols including, without limitation, direct connect, Ethernet (e.g., twisted pair 10/100/1000 Base T), Transmission Control Protocol/Internet Protocol (TCP/IP), token ring, IEEE 802.11a/b/g/n/x, etc. Using the network interface 403 and the communication network 409, the computer system 400 may communicate with gesture detection unit 116. The computer system 400 may include at least one of unit in the vehicle such as, but not limited to, Electronic/Engine Control Unit (ECU), Body Control Unit (BCU), Telematics/Transmission Control Unit (TCU), Door Control Unit (DCU), Human machine interface (HMI), Powertrain Control Module (PCM), Speed Control Unit (SCU), Batter Management System (BMS), Driver Monitoring System (DMS), infotainment system, and the like. The image capturing device 102 may communicate with the computer system 400. Further, the computer system 400 may communicate with the actuators 106. The actuators 106 may communicate with the one or more mirrors 108. The image capturing device 102 may include at least one of, but not limited to, a camera, a depth camera, an infrared camera, a smart camera, a 360-degree camera, a night vision camera, Red Green Blue (RGB) camera, three-dimensional (3D) camera, an eye recognition camera, a sensor, and the like.
[0091] The communication network 409 can be implemented as one of the different types of networks, such as intranet or Local Area Network (LAN) and such within the organization. The communication network 409 may either be a dedicated network or a shared network, which represents an association of the different types of networks that use a variety of protocols, for example, Hypertext Transfer Protocol (HTTP), Transmission Control Protocol/Internet Protocol (TCP/IP), Wireless Application Protocol (WAP), etc., to communicate with each other. Further, the communication network 409 may include a variety of network devices, including routers, bridges, servers, computing devices, storage devices, etc. In some embodiments, the processor 402 may be disposed in communication with a memory 405 (e.g., RAM, ROM, etc. not shown in FIG. 4) via a storage interface 404. The storage interface 404 may connect to memory 405 including, without limitation, memory drives, removable disc drives, etc., employing connection protocols such as Serial Advanced Technology Attachment (SATA), Integrated Drive Electronics (IDE), IEEE-1395, Universal Serial Bus (USB), fibre channel, Small Computer Systems Interface (SCSI), etc. The memory drives may further include a drum, magnetic disc drive, magneto-optical drive, optical drive, Redundant Array of Independent Discs (RAID), solid-state memory devices, solid-state drives, etc. The memory 405 may store a collection of program or database components, including, without limitation, a user interface 406, an operating system 407, a web browser 408, etc. In some embodiments, the computer system 400 may store user/application data, such as the data, variables, records, etc. as described in this invention. Such databases may be implemented as fault-tolerant, relational, scalable, secure databases such as Oracle or Sybase. Operating system 407 may facilitate resource management and operation of the computer system 400. Examples of operating systems include, without limitation, APPLE® MACINTOSH® OS X®, UNIX®, UNIX-like system distributions (E.G., BERKELEY SOFTWARE DISTRIBUTION® (BSD), FREEBSD®, NETBSD®, OPENBSD, etc.), LINUX® DISTRIBUTIONS (E.G., RED HAT®, UBUNTU®, KUBUNTU®, etc.), IBM®OS/2®, MICROSOFT® WINDOWS® (XP®, VISTA®/7/8, 10 etc.), APPLE® IOS®, GOOGLETM ANDROIDTM, BLACKBERRY® OS, or the like. User interface 406 may facilitate display, execution, interaction, manipulation, or operation of program components through textual or graphical facilities. For example, user interfaces may provide computer interaction interface elements on a display system operatively connected to computer system 400, such as cursors, icons, checkboxes, menus, scrollers, windows, widgets, etc. Graphical User Interfaces (GUIs) may be employed, including, without limitation, Apple® Macintosh® operating systems’ Aqua®, IBM® OS/2®, Microsoft® Windows® (e.g., Aero, Metro, etc.), web interface libraries (e.g., ActiveX®, Java®, Javascript®, AJAX, HTML, Adobe® Flash®, etc.), or the like.
[0092] Computer system 400 may implement a web browser 408 stored-program components. Web browser 408 may be a hypertext viewing application, such as MICROSOFT® INTERNET EXPLORER®, GOOGLETM CHROMETM, MOZILLA® FIREFOX®, APPLE® SAFARI®, etc. Secure web browsing may be provided using Secure Hypertext Transport Protocol (HTTPS), Secure Sockets Layer (SSL), Transport Layer Security (TLS), etc. Web browsers 408 may utilize facilities such as AJAX, DHTML, ADOBE® FLASH®, JAVASCRIPT®, JAVA®, Application Programming Interfaces (APIs), etc. Computer system 400 may implement a mail server stored program component. The mail server may be an Internet mail server such as Microsoft Exchange, or the like. The mail server may utilize facilities such as ASP, ACTIVEX®, ANSI® C++/C#, MICROSOFT®, NET, CGI SCRIPTS, JAVA®, JAVASCRIPT®, PERL®, PHP, PYTHON®, WEBOBJECTS®, etc. The mail server may utilize communication protocols such as Internet Message Access Protocol (IMAP), Messaging Application Programming Interface (MAPI), MICROSOFT® exchange, Post Office Protocol (POP), Simple Mail Transfer Protocol (SMTP), or the like. In some embodiments, the computer system 400 may implement a mail client stored program component. The mail client may be a mail viewing application, such as APPLE® MAIL, MICROSOFT® ENTOURAGE®, MICROSOFT® OUTLOOK®, MOZILLA® THUNDERBIRD®, etc. Furthermore, one or more computer-readable storage media may be utilized in implementing embodiments consistent with the present invention. A computer-readable storage medium refers to any type of physical memory on which information or data readable by a processor may be stored. Thus, a computer-readable storage medium may store instructions for execution by one or more processors, including instructions for causing the processor(s) to perform steps or stages consistent with the embodiments described herein. The term “computer-readable medium” should be understood to include tangible items and exclude carrier waves and transient signals, i.e., non-transitory. Examples include Random Access Memory (RAM), Read-Only Memory (ROM), volatile memory, non-volatile memory, hard drives, Compact Disc (CD) ROMs, Digital Video Disc (DVDs), flash drives, disks, and any other known physical storage media.
[0093] Embodiments herein enable automatic adjustment of one or more mirrors using eye point and hip point of the user driving the vehicle, which advantageously aids in providing appropriate field of view to the user without blind spot. By combining the eye point and hip point (height /position of the user), the exact distance from the eyes to the rear-view mirrors, an appropriate vertical and horizontal angle of the one or more mirrors can be determined, that will allow the user driving the vehicle to see proper rear view, while minimizing the size of the “blind spot” that can occur. Use of the mirror adjustment system disclosed in the embodiments herein aids in providing accuracy in positioning of the rear-view mirrors to reflect objects in the field of view and the appreciation by the user driving the vehicle for proper mirror settings are greatly increased. This results in eliminating manual adjustment of the mirror while driving or improper adjustments of the mirror, and in turn leading to lower frequency of accidents and injury. Since, the mirrors are appropriately adjusted to each user, there is more accuracy of mirror adjustment which seems appropriate to the field of view to the user and the seat position of the user. Further, the present disclosure provides a feature wherein the one or more mirrors of the vehicle are continuously adjusted in real-time or adjusted in real-time upon occurrence of events such as seat repositioning, change in driver etc., which not only completely eliminates manual adjustment of the mirrors, but also provides the flexibility of auto adapting the mirrors to changing environment such as various drivers, various seat positioning etc.
[0094] Embodiments herein allows the user to customize the gestures which is more suitable and easily remembered by the user to control the one or more systems in the vehicle. Thus, by performing the customized gestures to control the one or more systems, the user can concentrate on driving the vehicle more efficiently, in turn leading to lesser distraction, and lower frequency of accidents and injury.
[0095] A description of an embodiment with several components in communication with each other does not imply that all such components are required. On the contrary, a variety of optional components are described to illustrate the wide variety of possible embodiments of the invention. When a single device or article is described herein, it will be apparent that more than one device/article (whether or not they cooperate) may be used in place of a single device/article. Similarly, where more than one device or article is described herein (whether or not they cooperate), it will be apparent that a single device/article may be used in place of the more than one device or article or a different number of devices/articles may be used instead of the shown number of devices or programs. The functionality and/or the features of a device may be alternatively embodied by one or more other devices which are not explicitly described as having such functionality/features. Thus, other embodiments of the invention need not include the device itself.
[0096] The specification has described method and mirror adjustment system for dynamically adjusting one or more mirrors in a vehicle. The illustrated steps are set out to explain the exemplary embodiments shown, and it should be anticipated that ongoing technological development will change the manner in which particular functions are performed. These examples are presented herein for purposes of illustration, and not limitation. Further, the boundaries of the functional building blocks have been arbitrarily defined herein for the convenience of the description. Alternative boundaries can be defined so long as the specified functions and relationships thereof are appropriately performed. Alternatives (including equivalents, extensions, variations, deviations, etc., of those described herein) will be apparent to persons skilled in the relevant art(s) based on the teachings contained herein. Such alternatives fall within the scope and spirit of the disclosed embodiments. Also, the words "comprising," "having," "containing," and "including," and other similar forms are intended to be equivalent in meaning and be open-ended in that an item or items following any one of these words is not meant to be an exhaustive listing of such item or items or meant to be limited to only the listed item or items. It must also be noted that as used herein and in the appended claims, the singular forms “a,” “an,” and “the” include plural references unless the context clearly dictates otherwise.
[0097] Finally, the language used in the specification has been principally selected for readability and instructional purposes, and it may not have been selected to delineate or circumscribe the inventive subject matter. It is therefore intended that the scope of the invention be limited not by this detailed description, but rather by any claims that issue on an application based here on. Accordingly, the embodiments of the present invention are intended to be illustrative, but not limiting, of the scope of the invention, which is set forth in the following claims.
Referral numerals
Reference Number Description
100 Architecture
102 Image capturing device
104 Mirror adjustment system
106 Actuators
108 One or more mirrors
110 Processor
112 Input/Output (I/O) interface
114 Memory
116 Gesture detection unit
202 Data of mirror adjustment system
204 Modules
205 Data of gesture detection unit
206 Image data
208 Eye point data
210 Hip point data
212 Parametric data
214 Other data of mirror adjustment system
216 Gesture data
218 Action data
220 User profile data
221 Other data of gesture detection unit
222 Receiving module
224 Eye point determining module
226 Hip point determining module
228 Parameter determining module
230 Controlling module
232 Other modules
240 Vehicle
242 User
400 Computer system
401 I/O Interface of computer system
402 Processor of computer system
403 Network interface
404 Storage interface
405 Memory of computer system
406 User interface
407 Operating system
408 Web browser
409 Communication network
411 Input device
412 Output device
| # | Name | Date |
|---|---|---|
| 1 | 202121011870-STATEMENT OF UNDERTAKING (FORM 3) [19-03-2021(online)].pdf | 2021-03-19 |
| 2 | 202121011870-REQUEST FOR EXAMINATION (FORM-18) [19-03-2021(online)].pdf | 2021-03-19 |
| 3 | 202121011870-POWER OF AUTHORITY [19-03-2021(online)].pdf | 2021-03-19 |
| 4 | 202121011870-FORM 18 [19-03-2021(online)].pdf | 2021-03-19 |
| 5 | 202121011870-FORM 1 [19-03-2021(online)].pdf | 2021-03-19 |
| 6 | 202121011870-DRAWINGS [19-03-2021(online)].pdf | 2021-03-19 |
| 7 | 202121011870-DECLARATION OF INVENTORSHIP (FORM 5) [19-03-2021(online)].pdf | 2021-03-19 |
| 8 | 202121011870-COMPLETE SPECIFICATION [19-03-2021(online)].pdf | 2021-03-19 |
| 9 | 202121011870-FORM-8 [25-03-2021(online)].pdf | 2021-03-25 |
| 10 | Abstract1.jpg | 2021-10-19 |
| 11 | 202121011870-FER.pdf | 2022-10-17 |
| 12 | 202121011870-Proof of Right [25-01-2023(online)].pdf | 2023-01-25 |
| 13 | 202121011870-FER_SER_REPLY [20-02-2023(online)].pdf | 2023-02-20 |
| 14 | 202121011870-CLAIMS [20-02-2023(online)].pdf | 2023-02-20 |
| 15 | 202121011870-Response to office action [10-11-2023(online)].pdf | 2023-11-10 |
| 16 | 202121011870-FORM-26 [10-11-2023(online)].pdf | 2023-11-10 |
| 17 | 202121011870-PatentCertificate24-11-2023.pdf | 2023-11-24 |
| 18 | 202121011870-IntimationOfGrant24-11-2023.pdf | 2023-11-24 |
| 19 | 202121011870-POWER OF AUTHORITY [28-01-2025(online)].pdf | 2025-01-28 |
| 20 | 202121011870-FORM-16 [28-01-2025(online)].pdf | 2025-01-28 |
| 21 | 202121011870-ASSIGNMENT WITH VERIFIED COPY [28-01-2025(online)].pdf | 2025-01-28 |
| 1 | 202121011870searchstrategyE_14-10-2022.pdf |