Sign In to Follow Application
View All Documents & Correspondence

Method And System For Avoiding Collision Between Inspection Devices During Inspection Of An Area

Abstract: Disclosed herein is method and collision avoidance system for avoiding collision between inspection devices during inspection of an area. In an embodiment, spatial information of the area, dynamically received from each inspection device, is used for creating current local map for each inspection device. Subsequently, a global map of area is generated using local maps and each inspection device is localized within the global map for identifying current location of the inspection devices. Thereafter, sub-paths for navigating each inspection device are estimated based on global map and current location of the inspection devices, validated for collision avoidance and provided to corresponding inspection devices. The inspection devices navigate across the area using the sub-paths and complete inspection of the area. In some embodiments, the present disclosure helps in faster and accurate inspection of the area by inspecting the area using concurrently operated inspection devices. FIG. 1

Get Free WhatsApp Updates!
Notices, Deadlines & Correspondence

Patent Information

Application #
Filing Date
29 March 2019
Publication Number
40/2020
Publication Type
INA
Invention Field
COMPUTER SCIENCE
Status
Email
bangalore@knspartners.com
Parent Application
Patent Number
Legal Status
Grant Date
2025-02-07
Renewal Date

Applicants

WIPRO LIMITED
Doddakannelli, Sarjapur Road, Bangalore 560035, Karnataka, India.

Inventors

1. NIKHIL PAREEK
SJR Verity, VIDI-503, Kasavanahalli, Jail Road, Bangalore 560035, Karnataka, India.
2. VINDOD PATHANGAY
E-1902, Ajmera Infinity, Neeladri Road, Electronics City Phase-1, Bangalore 560100

Specification

Claims:WE CLAIM:
1. A method of avoiding collision between inspection devices (101) during inspection of an area, the method comprising:
receiving, by a collision avoidance system (105), spatial information (211) of the area from each of a plurality of inspection devices (101), initialized for inspecting the area, wherein each of the plurality of inspection devices (101) navigate across a plurality of hops and the spatial information (211) is received at beginning of each of the plurality of hops;
creating, by the collision avoidance system (105), a current local map (213) for each of the plurality of inspection devices (101) based on the spatial information (211) of a current hop of the plurality of hops of each of the plurality of inspection devices (101);
generating, by the collision avoidance system (105), a current global map (217) of the area by merging the current local map (213) of each of the plurality of inspection devices (101) with a previous global map (215) of the area, wherein the previous global map (215) corresponds to previous hops of the plurality of hops of each of the plurality of inspection devices (101);
localizing, by the collision avoidance system (105), each of the plurality of inspection devices (101) within the current global map (217) for determining a current location of each of the plurality of inspection devices (101);
estimating, by the collision avoidance system (105), one or more sub-paths (219) for navigating each of the plurality of inspection devices (101) from the current hop to a subsequent hop of the plurality of hops, by analysing the current global map (217) and the current location of each of the plurality of inspection devices (101) using a pre-trained path estimation model (229), wherein each of the one or more sub-paths (219) are validated for eliminating possibility of collisions among the plurality of inspection devices (101) in the subsequent hop; and
providing, by the collision avoidance system (105), each of the one or more sub-paths (219) to each of the plurality of inspection devices (101) for avoiding the collision between the plurality of inspection devices (101) during inspection of the area.

2. The method as claimed in claim 1, wherein the spatial information (211) is collected using one or more sensors (307) and image capturing devices (309) configured with each of the plurality of inspection devices (101).

3. The method as claimed in claim 1, wherein generating the current global map (217) of the area comprises:
identifying one or more adjacent local maps based on comparison among the current local map (213) corresponding to each of the plurality of inspection devices (101);
assigning a continuity score for each of the one or more adjacent local maps based on visual similarity among the one or more adjacent local maps; and
merging the one or more adjacent local maps into the current global map (217), when the continuity score of the one or more adjacent local maps is greater than a threshold score.

4. The method as claimed in claim 1, wherein validating each of the one or more sub-paths (219) comprises:
estimating a future location of the plurality of inspection devices (101) based on analysis of the current location of the plurality of inspection devices (101) and the one or more sub-paths (219);
determining possibility of collision between the plurality of inspection devices (101) at the future location of the plurality of inspection devices (101), using a pre-trained reinforcement model; and
eliminating the one or more sub-paths (219) having the possibility of collision.

5. The method as claimed in claim 1, wherein inspecting the area comprises:
capturing, using the plurality of inspection devices (101), a plurality of images of a region to be inspected within the area;
ranking each of the plurality of images based on one or more objects present in the plurality of images, wherein the ranking is performed using a pre-trained machine learning model;
combining one or more of the plurality of images into one or more inspection images based on the ranking; and
displaying the one or more inspection images on a display unit (206) associated with the collision avoidance system (105) for inspecting the area.

6. A collision avoidance system (105) for avoiding collision between inspection devices (101) during inspection of an area, the collision avoidance system (105) comprising:
a processor (203); and
a memory (205), communicatively coupled to the processor (203), wherein the memory (205) stores processor-executable instructions, which on execution, cause the processor (203) to:
receive spatial information (211) of the area from each of a plurality of inspection devices (101), initialized for inspecting the area, wherein each of the plurality of inspection devices (101) navigate across a plurality of hops and the spatial information (211) is received at beginning of each of the plurality of hops;
create a current local map (213) for each of the plurality of inspection devices (101) based on the spatial information (211) of a current hop of the plurality of hops of the each of the plurality of inspection devices (101);
generate a current global map (217) of the area by merging the current local map (213) of each of the plurality of inspection devices (101) with a previous global map (215) of the area, wherein the previous global map (215) corresponds to previous hops of the plurality of hops of each of the plurality of inspection devices (101);
localize each of the plurality of inspection devices (101) within the current global map (217) for determining a current location of each of the plurality of inspection devices (101);
estimate one or more sub-paths (219) for navigating each of the plurality of inspection devices (101) from the current hop to a subsequent hop of the plurality of hops, by analysing the current global map (217) and the current location of each of the plurality of inspection devices (101) using a pre-trained path estimation model (229), wherein each of the one or more sub-paths (219) are validated for eliminating possibility of collisions among the plurality of inspection devices (101) in the subsequent hop; and
provide each of the one or more sub-paths (219) to each of the plurality of inspection devices (101) for avoiding the collision between the plurality of inspection devices (101) during inspection of the area.

7. The collision avoidance system (105) as claimed in claim 6, wherein each of the plurality of inspection devices (101) collect the spatial information (211) using one or more sensors (307) and image capturing devices (309) configured with each of the plurality of inspection devices (101).

8. The collision avoidance system (105) as claimed in claim 6, wherein to generate the current global map (217) of the area, the processor (203) is configured to:
identify one or more adjacent local maps based on comparison among the current local map (213) corresponding to each of the plurality of inspection devices (101);
assign a continuity score for each of the one or more adjacent local maps based on visual similarity among the one or more adjacent local maps; and
merge the one or more adjacent local maps into the current global map (217), when the continuity score of the one or more adjacent local maps is greater than a threshold score.

9. The collision avoidance system (105) as claimed in claim 6, wherein to validate each of the one or more sub-paths (219), the processor (203) is configured to:
estimate a future location of the plurality of inspection devices (101) based on analysis of the current location of the plurality of inspection devices (101) and the one or more sub-paths (219);
determine possibility of collision between the plurality of inspection devices (101) at the future location of the plurality of inspection devices (101), using a pre-trained reinforcement model; and
eliminate the one or more sub-paths (219) having the possibility of collision.

10. The collision avoidance system (105) as claimed in claim 6, wherein to inspect the area, the processor (203) is configured to:
capture, using the plurality of inspection devices (101), a plurality of images of a region to be inspected within the area;
rank each of the plurality of images based on one or more objects present in the plurality of images, wherein the ranking is performed using a pre-trained machine learning model;
combine one or more of the plurality of images into one or more inspection images based on the ranking; and
display the one or more inspection images on a display unit (206) associated with the collision avoidance system (105) for inspecting the area.

Dated this 29th day of March 2019

MADHUSUDAN S.T
OF K&S PARTNERS
ATTORNEY FOR THE APPLICANT
IN/PA-1297
, Description:TECHNICAL FIELD
The present subject matter is, in general, related to automated inspection and more particularly, but not exclusively, to method and system for avoiding collision between inspection devices during inspection of an area.

Documents

Orders

Section Controller Decision Date
15 and 43 PRANAV KUMAR 2025-02-07
15 and 43 PRANAV KUMAR 2025-02-07

Application Documents

# Name Date
1 201941012478-IntimationOfGrant07-02-2025.pdf 2025-02-07
1 201941012478-Request Letter-Correspondence [04-02-2019(online)].pdf 2019-02-04
2 201941012478-PatentCertificate07-02-2025.pdf 2025-02-07
2 201941012478-Power of Attorney [04-02-2019(online)].pdf 2019-02-04
3 201941012478-FORM-26 [22-02-2024(online)].pdf 2024-02-22
3 201941012478-Form 1 (Submitted on date of filing) [04-02-2019(online)].pdf 2019-02-04
4 201941012478-Written submissions and relevant documents [22-02-2024(online)].pdf 2024-02-22
4 201941012478-STATEMENT OF UNDERTAKING (FORM 3) [29-03-2019(online)].pdf 2019-03-29
5 201941012478-REQUEST FOR EXAMINATION (FORM-18) [29-03-2019(online)].pdf 2019-03-29
5 201941012478-AMENDED DOCUMENTS [18-01-2024(online)].pdf 2024-01-18
6 201941012478-POWER OF AUTHORITY [29-03-2019(online)].pdf 2019-03-29
6 201941012478-Correspondence to notify the Controller [18-01-2024(online)].pdf 2024-01-18
7 201941012478-FORM 18 [29-03-2019(online)].pdf 2019-03-29
7 201941012478-FORM 13 [18-01-2024(online)].pdf 2024-01-18
8 201941012478-POA [18-01-2024(online)].pdf 2024-01-18
8 201941012478-FORM 1 [29-03-2019(online)].pdf 2019-03-29
9 201941012478-DRAWINGS [29-03-2019(online)].pdf 2019-03-29
9 201941012478-US(14)-HearingNotice-(HearingDate-07-02-2024).pdf 2024-01-04
10 201941012478-DECLARATION OF INVENTORSHIP (FORM 5) [29-03-2019(online)].pdf 2019-03-29
10 201941012478-FER.pdf 2021-10-17
11 201941012478-CLAIMS [20-09-2021(online)].pdf 2021-09-20
11 201941012478-COMPLETE SPECIFICATION [29-03-2019(online)].pdf 2019-03-29
12 201941012478-COMPLETE SPECIFICATION [20-09-2021(online)].pdf 2021-09-20
12 201941012478-Proof of Right (MANDATORY) [24-09-2019(online)].pdf 2019-09-24
13 201941012478-DRAWING [20-09-2021(online)].pdf 2021-09-20
13 Correspondence by Agent_Form-1_03-10-2019.pdf 2019-10-03
14 201941012478-FER_SER_REPLY [20-09-2021(online)].pdf 2021-09-20
14 201941012478-FORM 3 [19-09-2021(online)].pdf 2021-09-19
15 201941012478-PETITION UNDER RULE 137 [20-09-2021(online)].pdf 2021-09-20
16 201941012478-FER_SER_REPLY [20-09-2021(online)].pdf 2021-09-20
16 201941012478-FORM 3 [19-09-2021(online)].pdf 2021-09-19
17 Correspondence by Agent_Form-1_03-10-2019.pdf 2019-10-03
17 201941012478-DRAWING [20-09-2021(online)].pdf 2021-09-20
18 201941012478-Proof of Right (MANDATORY) [24-09-2019(online)].pdf 2019-09-24
18 201941012478-COMPLETE SPECIFICATION [20-09-2021(online)].pdf 2021-09-20
19 201941012478-CLAIMS [20-09-2021(online)].pdf 2021-09-20
19 201941012478-COMPLETE SPECIFICATION [29-03-2019(online)].pdf 2019-03-29
20 201941012478-DECLARATION OF INVENTORSHIP (FORM 5) [29-03-2019(online)].pdf 2019-03-29
20 201941012478-FER.pdf 2021-10-17
21 201941012478-DRAWINGS [29-03-2019(online)].pdf 2019-03-29
21 201941012478-US(14)-HearingNotice-(HearingDate-07-02-2024).pdf 2024-01-04
22 201941012478-FORM 1 [29-03-2019(online)].pdf 2019-03-29
22 201941012478-POA [18-01-2024(online)].pdf 2024-01-18
23 201941012478-FORM 13 [18-01-2024(online)].pdf 2024-01-18
23 201941012478-FORM 18 [29-03-2019(online)].pdf 2019-03-29
24 201941012478-Correspondence to notify the Controller [18-01-2024(online)].pdf 2024-01-18
24 201941012478-POWER OF AUTHORITY [29-03-2019(online)].pdf 2019-03-29
25 201941012478-REQUEST FOR EXAMINATION (FORM-18) [29-03-2019(online)].pdf 2019-03-29
25 201941012478-AMENDED DOCUMENTS [18-01-2024(online)].pdf 2024-01-18
26 201941012478-Written submissions and relevant documents [22-02-2024(online)].pdf 2024-02-22
26 201941012478-STATEMENT OF UNDERTAKING (FORM 3) [29-03-2019(online)].pdf 2019-03-29
27 201941012478-FORM-26 [22-02-2024(online)].pdf 2024-02-22
27 201941012478-Form 1 (Submitted on date of filing) [04-02-2019(online)].pdf 2019-02-04
28 201941012478-Power of Attorney [04-02-2019(online)].pdf 2019-02-04
28 201941012478-PatentCertificate07-02-2025.pdf 2025-02-07
29 201941012478-Request Letter-Correspondence [04-02-2019(online)].pdf 2019-02-04
29 201941012478-IntimationOfGrant07-02-2025.pdf 2025-02-07

Search Strategy

1 robot-operating-system-ros-2017AE_28-12-2021.pdf
2 collisionmarketrobotE_16-03-2021.pdf
3 2021-03-1614-51-01E_16-03-2021.pdf

ERegister / Renewals

3rd: 06 May 2025

From 29/03/2021 - To 29/03/2022

4th: 06 May 2025

From 29/03/2022 - To 29/03/2023

5th: 06 May 2025

From 29/03/2023 - To 29/03/2024

6th: 06 May 2025

From 29/03/2024 - To 29/03/2025

7th: 06 May 2025

From 29/03/2025 - To 29/03/2026