Abstract: This disclosure relates to method and system for providing an explanation for a prediction generated by an artificial neural network (ANN) model for a given input data. The method may include receiving the given input data and the prediction generated by the ANN model. The ANN model may be built and trained for a target application. The method may further include determining a plurality of relevant portions of the given input data. For each of the plurality of relevant portions, the method may further include fetching a portional prediction and a portional prediction score generated by the ANN model, and determining a degree of influence score based on the portional prediction score and a comparison between the portional prediction and the prediction. The method may further include providing the explanation for the prediction based on the degree of influence score of each of the plurality of relevant portions. Figure 6
Claims:WE CLAIM
1. A method of providing an explanation for a prediction generated by an artificial neural network (ANN) model for a given input data, the method comprising:
receiving, by a prediction explanation device, the given input data and the prediction generated by the ANN model, wherein the ANN model is built and trained for a target application;
determining, by the prediction explanation device, a plurality of relevant portions of the given input data;
for each of the plurality of relevant portions,
fetching, by the prediction explanation device, a portional prediction and a portional prediction score generated by the ANN model; and
determining, by the prediction explanation device, a degree of influence score based on the portional prediction score and a comparison between the portional prediction and the prediction; and
providing, by the prediction explanation device, the explanation for the prediction based on the degree of influence score of each of the plurality of relevant portions.
2. The method of claim 1, wherein the given input data comprises at least one of text data, audio data, video data, and image data.
3. The method of claim 1, wherein determining the plurality of relevant portions comprises:
segmenting the given input data into a plurality of portions; and
processing each of the plurality of portions to filter the plurality of relevant portions.
4. The method of claim 1, wherein the target application comprises a text based application, wherein the ANN model comprises a recurrent neural network (RNN) model, and wherein each of the plurality of relevant portions comprises a relevant token from a tokenized text.
5. The method of claim 1, wherein the providing the explanations further comprises determining a set of influential portions from among the plurality of relevant portions based on the degree of influence score of each of the plurality of relevant portions.
6. The method of claim 5, further comprising retuning the ANN model based on the set of influential portions.
7. The method of claim 1, wherein the providing the explanations comprises rendering each of the plurality of relevant portions along with the corresponding degree of influence score.
8. A system for providing an explanation for a prediction generated by an artificial neural network (ANN) model for a given input data, the system comprising:
a prediction explanation device comprising at least one processor and a computer-readable medium storing instructions that, when executed by the at least one processor, cause the at least one processor to perform operations comprising:
receiving the given input data and the prediction generated by the ANN model, wherein the ANN model is built and trained for a target application;
determining a plurality of relevant portions of the given input data;
for each of the plurality of relevant portions,
fetching a portional prediction and a portional prediction score generated by the ANN model; and
determining a degree of influence score based on the portional prediction score and a comparison between the portional prediction and the prediction; and
providing the explanation for the prediction based on the degree of influence score of each of the plurality of relevant portions.
9. The system of claim 8, wherein the given input data comprises at least one of text data, audio data, video data, and image data.
10. The system of claim 8, wherein determining the plurality of relevant portions comprises:
segmenting the given input data into a plurality of portions; and
processing each of the plurality of portions to filter the plurality of relevant portions.
11. The system of claim 8, wherein the target application comprises a text based application, wherein the ANN model comprises a recurrent neural network (RNN) model, and wherein each of the plurality of relevant portions compres a relevant token from a tokenized text.
12. The system of claim 8, wherein the providing the explanations further comprises determining a set of influential portions from among the plurality of relevant portions based on the degree of influence score of each of the plurality of relevant portions.
13. The system of claim 12, wherein the operations further comprise retuning the ANN model based on the set of influential portions.
14. The system of claim 8, wherein the providing the explanations comprises rendering each of the plurality of relevant portions along with the corresponding degree of influence score.
Dated this 31st day of December, 2018
Madhusudan S.T.
Of K&S Partners
Agent for the Applicant
IN/PA-1297
, Description:TECHNICAL FIELD
This disclosure relates generally to artificial neural network (ANN), and more particularly to method and system for providing an explanation for a prediction generated by an ANN model.
| # | Name | Date |
|---|---|---|
| 1 | 201841049976-STATEMENT OF UNDERTAKING (FORM 3) [31-12-2018(online)].pdf | 2018-12-31 |
| 2 | 201841049976-REQUEST FOR EXAMINATION (FORM-18) [31-12-2018(online)].pdf | 2018-12-31 |
| 3 | 201841049976-POWER OF AUTHORITY [31-12-2018(online)].pdf | 2018-12-31 |
| 4 | 201841049976-FORM 18 [31-12-2018(online)].pdf | 2018-12-31 |
| 5 | 201841049976-FORM 1 [31-12-2018(online)].pdf | 2018-12-31 |
| 6 | 201841049976-DRAWINGS [31-12-2018(online)].pdf | 2018-12-31 |
| 7 | 201841049976-DECLARATION OF INVENTORSHIP (FORM 5) [31-12-2018(online)].pdf | 2018-12-31 |
| 8 | 201841049976-COMPLETE SPECIFICATION [31-12-2018(online)].pdf | 2018-12-31 |
| 9 | Abstract_201841049976.jpg | 2019-01-03 |
| 10 | 201841049976-Request Letter-Correspondence [03-01-2019(online)].pdf | 2019-01-03 |
| 11 | 201841049976-Power of Attorney [03-01-2019(online)].pdf | 2019-01-03 |
| 12 | 201841049976-Form 1 (Submitted on date of filing) [03-01-2019(online)].pdf | 2019-01-03 |
| 13 | 201841049976-Proof of Right (MANDATORY) [13-06-2019(online)].pdf | 2019-06-13 |
| 14 | Correspondence by Agent_Form1_20-06-2019.pdf | 2019-06-20 |
| 15 | 201841049976-FER.pdf | 2021-10-17 |
| 16 | 201841049976-POA [28-02-2022(online)].pdf | 2022-02-28 |
| 17 | 201841049976-OTHERS [28-02-2022(online)].pdf | 2022-02-28 |
| 18 | 201841049976-FORM 13 [28-02-2022(online)].pdf | 2022-02-28 |
| 19 | 201841049976-FER_SER_REPLY [28-02-2022(online)].pdf | 2022-02-28 |
| 20 | 201841049976-DRAWING [28-02-2022(online)].pdf | 2022-02-28 |
| 21 | 201841049976-CLAIMS [28-02-2022(online)].pdf | 2022-02-28 |
| 22 | 201841049976-AMENDED DOCUMENTS [28-02-2022(online)].pdf | 2022-02-28 |
| 23 | 201841049976-PETITION UNDER RULE 137 [02-03-2022(online)].pdf | 2022-03-02 |
| 24 | 201841049976-PatentCertificate17-11-2023.pdf | 2023-11-17 |
| 25 | 201841049976-IntimationOfGrant17-11-2023.pdf | 2023-11-17 |
| 1 | SearchStrategyMatrixE_09-07-2021.pdf |