Sign In to Follow Application
View All Documents & Correspondence

System And Method For Implementing Neural Network Models On Edge Devices In Lot Networks

Abstract: A method and a system for implementing neural network models on edge devices in an Internet of Things (IoT) network are disclosed. In an embodiment, the method may include receiving a neural network model trained and configured to detect objects from images, and iteratively assigning a new value to each of a plurality of parameters associated with the neural network model to generate a re-configured neural network model in each iteration. The method may further include deploying for a current iteration the re-configured neural network on the edge device. The method may further include computing for the current iteration, a trade-off value based on a detection accuracy associated with the at least one object detected in the image and resource utilization data associated with the edge device, and selecting the re-configured neural network model, based on the trade-off value calculated for the current iteration. Fig. 1

Get Free WhatsApp Updates!
Notices, Deadlines & Correspondence

Patent Information

Application #
Filing Date
03 June 2019
Publication Number
49/2020
Publication Type
INA
Invention Field
COMPUTER SCIENCE
Status
Email
ipr@akshipassociates.com
Parent Application
Patent Number
Legal Status
Grant Date
2023-10-18
Renewal Date

Applicants

WIPRO LIMITED
Doddakannelli, Sarjapur Road, Bangalore 560035, Karnataka, India.

Inventors

1. NIDHI MITTAL HADA
50 A Turf Road. Bhowanipore, Kolkata, West Bengal, India
2. DEBASISH CHANDA
Daulatpur, Phoolbagan, 24 Pargans (S) 700140, West Bengal, India

Specification

Claims:We claim:
1. A method of implementing neural network models on edge devices in an Internet of Things (IoT) network, the method comprising:
receiving, by an optimizing device, a neural network model trained and configured to detect objects from images, wherein the neural network model is a Convolutional Neural Network (CNN);
iteratively assigning, by the optimizing device, a new value to each of a plurality of parameters associated with the neural network model to generate a re-configured neural network model in each iteration;
deploying for a current iteration, by the optimizing device, the re-configured neural network on an edge device to identify at least one object in an image;
computing, by the optimizing device, for the current iteration, a trade-off value based on a detection accuracy associated with the at least one object detected in the image and resource utilization data associated with the edge device; and
selecting, by the optimizing device, the re-configured neural network model of the current iteration, based on the trade-off value calculated for the current iteration.

2. The method of claim 1, wherein the selecting further comprises:
comparing the trade-off value calculated for the current iteration with a predefined threshold; and
selecting the re-configured neural network model of the current iteration, when the trade-off value is greater than or equal to the predefined threshold.

3. The method of claim 1 further comprising:
rendering a parameter interface to a user, wherein the parameter interface is configured to receive values assigned to one or more of the plurality of parameters from the user, wherein the values assigned to the one or more of the plurality of parameters is used for iteratively assigning the new value to each of a plurality of parameters associated with the neural network model.

4. The method of claim 1, wherein the resource usage data associated with the edge device corresponds to one or more of a loss function, memory usage, and Millions of Instructions Per Second (MIPS) usage of the re-configured neural network model for the edge device.

5. The method of claim 1, wherein the plurality of parameters comprises at least one of batch_normalize, a filter, a size, a stride, a pad, an activation, a maxpool_size, a maxpool_stride, a bit-width, a memory allocation, a batchsize, an input image resolution, an input image height, an input image width, a channel, a learning rate, an epoch, a decay of learning rate, a quantization of weights and biases, a gridsize, a bounding_box_ per_grid_cell, an output_class, and a weight for Loss function.

6. An optimizing device for implementing neural network models on edge devices in an Internet of Things (IoT) network, the optimizing device comprising:
a processor; and
a memory communicatively coupled to the processor, wherein the memory stores processor instructions, which, on execution, causes the processor to:
receive a neural network model trained and configured to detect objects from images, wherein the neural network model is a Convolutional Neural Network (CNN);
iteratively assign a new value to each of a plurality of parameters associated with the neural network model to generate a re-configured neural network model in each iteration;
deploy, for a current iteration, the re-configured neural network on an edge device to identify at least one object in an image;
compute for the current iteration, a trade-off value based on a detection accuracy associated with the at least one object detected in the image and resource utilization data associated with the edge device; and
select the re-configured neural network model of the current iteration, based on the trade-off value calculated for the current iteration.

7. The optimizing device of claim 6, wherein the processor instructions further cause the processor to:
receive an image from an image capturing device within the edge device; and
detect one or more objects in the received image, using the re-configured neural network model of the current iteration.

8. The optimizing device of claim 6, wherein the selecting further comprises:
comparing the trade-off value calculated for the current iteration with a predefined threshold; and
selecting the re-configured neural network model of the current iteration, when the trade-off value is greater than or equal to the predefined threshold.

9. The optimizing device of claim 6, wherein the processor instructions further cause the processor to render a parameter interface to a user, wherein the parameter interface is configured to receive values assigned to one or more of the plurality of parameters from the user.

10. The optimizing device of claim 9, wherein the values assigned to the one or more of the plurality of parameters is used for iteratively assigning the new value to each of a plurality of parameters associated with the neural network model.

Dated this 3rd day of June, 2019

R Ramya Rao
Of K&S Partners
Agent for the Applicant
IN/PA-1607
, Description:Technical Field
[001] This disclosure relates generally to implementing neural network models, and more particularly to a method and system of implementing neural network models on edge devices in Internet of Things (IoT) networks.

Documents

Application Documents

# Name Date
1 201941022027-IntimationOfGrant18-10-2023.pdf 2023-10-18
1 201941022027-STATEMENT OF UNDERTAKING (FORM 3) [03-06-2019(online)].pdf 2019-06-03
2 201941022027-REQUEST FOR EXAMINATION (FORM-18) [03-06-2019(online)].pdf 2019-06-03
2 201941022027-PatentCertificate18-10-2023.pdf 2023-10-18
3 201941022027-POWER OF AUTHORITY [03-06-2019(online)].pdf 2019-06-03
3 201941022027-CLAIMS [06-11-2021(online)].pdf 2021-11-06
4 201941022027-FORM 18 [03-06-2019(online)].pdf 2019-06-03
4 201941022027-FER_SER_REPLY [06-11-2021(online)].pdf 2021-11-06
5 201941022027-OTHERS [06-11-2021(online)].pdf 2021-11-06
5 201941022027-FORM 1 [03-06-2019(online)].pdf 2019-06-03
6 201941022027-PETITION UNDER RULE 137 [06-11-2021(online)].pdf 2021-11-06
6 201941022027-DRAWINGS [03-06-2019(online)].pdf 2019-06-03
7 201941022027-RELEVANT DOCUMENTS [06-11-2021(online)].pdf 2021-11-06
7 201941022027-DECLARATION OF INVENTORSHIP (FORM 5) [03-06-2019(online)].pdf 2019-06-03
8 201941022027-COMPLETE SPECIFICATION [03-06-2019(online)].pdf 2019-06-03
8 201941022027-AMENDED DOCUMENTS [27-10-2021(online)].pdf 2021-10-27
9 201941022027-Request Letter-Correspondence [04-06-2019(online)].pdf 2019-06-04
9 201941022027-FORM 13 [27-10-2021(online)].pdf 2021-10-27
10 201941022027-POA [27-10-2021(online)].pdf 2021-10-27
10 201941022027-Power of Attorney [04-06-2019(online)].pdf 2019-06-04
11 201941022027-FER.pdf 2021-10-17
11 201941022027-Form 1 (Submitted on date of filing) [04-06-2019(online)].pdf 2019-06-04
12 201941022027-Proof of Right (MANDATORY) [18-11-2019(online)].pdf 2019-11-18
13 201941022027-FER.pdf 2021-10-17
13 201941022027-Form 1 (Submitted on date of filing) [04-06-2019(online)].pdf 2019-06-04
14 201941022027-POA [27-10-2021(online)].pdf 2021-10-27
14 201941022027-Power of Attorney [04-06-2019(online)].pdf 2019-06-04
15 201941022027-FORM 13 [27-10-2021(online)].pdf 2021-10-27
15 201941022027-Request Letter-Correspondence [04-06-2019(online)].pdf 2019-06-04
16 201941022027-AMENDED DOCUMENTS [27-10-2021(online)].pdf 2021-10-27
16 201941022027-COMPLETE SPECIFICATION [03-06-2019(online)].pdf 2019-06-03
17 201941022027-DECLARATION OF INVENTORSHIP (FORM 5) [03-06-2019(online)].pdf 2019-06-03
17 201941022027-RELEVANT DOCUMENTS [06-11-2021(online)].pdf 2021-11-06
18 201941022027-DRAWINGS [03-06-2019(online)].pdf 2019-06-03
18 201941022027-PETITION UNDER RULE 137 [06-11-2021(online)].pdf 2021-11-06
19 201941022027-FORM 1 [03-06-2019(online)].pdf 2019-06-03
19 201941022027-OTHERS [06-11-2021(online)].pdf 2021-11-06
20 201941022027-FORM 18 [03-06-2019(online)].pdf 2019-06-03
20 201941022027-FER_SER_REPLY [06-11-2021(online)].pdf 2021-11-06
21 201941022027-POWER OF AUTHORITY [03-06-2019(online)].pdf 2019-06-03
21 201941022027-CLAIMS [06-11-2021(online)].pdf 2021-11-06
22 201941022027-REQUEST FOR EXAMINATION (FORM-18) [03-06-2019(online)].pdf 2019-06-03
22 201941022027-PatentCertificate18-10-2023.pdf 2023-10-18
23 201941022027-STATEMENT OF UNDERTAKING (FORM 3) [03-06-2019(online)].pdf 2019-06-03
23 201941022027-IntimationOfGrant18-10-2023.pdf 2023-10-18

Search Strategy

1 SearchStrategyMatrixE_13-04-2021.pdf

ERegister / Renewals

3rd: 18 Jan 2024

From 03/06/2021 - To 03/06/2022

4th: 18 Jan 2024

From 03/06/2022 - To 03/06/2023

5th: 18 Jan 2024

From 03/06/2023 - To 03/06/2024

6th: 01 Jun 2024

From 03/06/2024 - To 03/06/2025

7th: 02 Jun 2025

From 03/06/2025 - To 03/06/2026