Sign In to Follow Application
View All Documents & Correspondence

Method And System For Determining Classification Of Text

Abstract: Embodiments of the present disclosure discloses method and system for determining classification of text. The present disclosure discloses to receive text from plurality of texts and generating a pair of vector representation of the text using trained model parameters of a pair of LSTM units. The trained model parameters are obtained based on training of classification system using plurality of similar pair of texts and plurality of dissimilar pair of texts from the plurality of texts. Further, pair of vector representations are combined using a combiner operator to obtain a combined vector representation. The combiner operator is selected from a plurality of combiner operators based on the training using accuracy of classifier of classification system. The combined vector representation is provided to the classifier for determining classification of text. The present disclosure enhances the performance and generalisation of a classifier in cases of a multi-class classification. Figure 3

Get Free WhatsApp Updates!
Notices, Deadlines & Correspondence

Patent Information

Application #
Filing Date
17 February 2017
Publication Number
34/2018
Publication Type
INA
Invention Field
COMPUTER SCIENCE
Status
Email
ipr@akshipassociates.com
Parent Application
Patent Number
Legal Status
Grant Date
2023-01-18
Renewal Date

Applicants

WIPRO LIMITED
Doddakannelli, Sarjapur Road, Bangalore 560035, Karnataka, India.

Inventors

1. DEEPAK BHATT
D-302, Sriram Signia, Opp. Wipro Gate 13, Neeladri Road, Electronics City, Bangalore 560100, Karnataka, India.
2. PRASHANT SINGH
Flat 501, Site 105, Modini Blooming, Neeladri Nagar, 14th Cross Electronic City Phase 1, Bangalore 560100, Karnataka, India

Specification

Claims:We claim:
1. A method for determining classification of text, comprising:
receiving, by a text classification system (101), a text from a plurality of texts (212);
generating, by the text classification system (101), a first vector representation (213) of the text using first trained model parameters (219) of a first Long Short Term Memory (LSTM) unit 601 and a second vector representation (213) of the text using second trained model parameters (219) of a second LSTM unit (602), wherein the first and second trained model parameters (219) are obtained based on training of the text classification system (101) using a plurality of similar pair of texts and a plurality of dissimilar pair of texts from the plurality of texts (212);
combining, by the text classification system (101), the first vector representation (213) and the second vector representation (213) using a combiner operator (222) to obtain a combined vector representation (214), wherein the combiner operator (222) is selected from a plurality of combiner operators (222) based on the training using accuracy (221) of a classifier (104) of the text classification system (101); and
providing, by the text classification system (101), the combined vector representation (214) to the classifier (104) for determining a classification (215) of the text.

2. The method as claimed in claim 1 and further comprising:
cleaning, by the text classification system (101), the text to remove irrelevant data from the text.

3. The method as claimed in claim 1, wherein the training using the similar pair of texts and the dissimilar pair of texts comprises:
receiving, by the text classification system (101), a first text and a second text, wherein the first text and the second text are one of the plurality of similar pair of texts and one of the plurality of dissimilar pair of texts;
generating, by the text classification system (101), a first labeled text (216) for the first text and a second labeled text (216) for the second text based on a plurality of labels (217) retrieved from data source (105) associated with the text classification system (101);
generating, by the text classification system (101), a third vector representation (213) of the first text using first model parameters (218) of the first LSTM unit (601) and a fourth vector representation (218) of the second text using second model parameters (218) of the second LSTM unit (602); and
modifying, by the text classification system (101), the first model parameters (218) and the second model parameters (218) to obtain the first trained model parameters (219) and the second trained model parameters (219) respectively based on norm distance (220) associated between the third vector representation (213) and the fourth vector representation (213).

4. The method as claimed in claim 3, wherein the first labeled text (216) comprises the first text and one of the plurality of labels (217) associated with the first text and the second labeled text (216) comprises the second text and one of the plurality of labels (217) associated with the second text.

5. The method as claimed in claim 3, wherein modifying the first model parameters (218) and the second model parameters (218) comprises proximating exponential of the negative of norm distance (220) to value zero for the dissimilar pair of texts and proximating the exponential of the negative of norm distance (220) to value one for the similar pair of texts.

6. The method as claimed in claim 1, wherein selecting the combiner operator 222 comprises:
receiving, by the text classification system (101), each of the plurality of texts (212);
generating, by the text classification system (101), a fifth vector representation (213) for each of the plurality of texts (212) using the first trained model parameters (219) of the first LSTM unit 601 and a sixth vector representation (213) for each of the plurality of texts (212) using the second trained model parameters (219) of the second LSTM unit (602);
combining, by the text classification system, the fifth vector representation (213) and the sixth vector representation (213), for each of the plurality of texts (212), using each of the plurality of combiner operators (222) to obtain corresponding a plurality of combined vector representations (214);
providing, by the text classification system, each of the plurality of combined vector representations (214) to the classifier (104) to obtain classification (215) for each of the plurality of combined vector representations (214);
determining, by the text classification system, accuracy (221) for each of the classification (215); and
selecting, by the text classification system, the combiner operator (222), for each of the plurality of texts (212), from the plurality of combiner operators (222), based on the accuracy (221) of the classification (215) of the plurality of combined vector representations (214).

7. A text classification system (101) for determining classification of text, comprises:
a processor (106); and
a memory (109) communicatively coupled to the processor (106), wherein the memory (109) stores processor-executable instructions, which, on execution, cause the processor to:
receive a text from a plurality of texts (212);
generate a first vector representation (213) of the text using first trained model parameters (219) of a first Long Short Term Memory (LSTM) unit 601 and a second vector representation (213) of the text using second trained model parameters (219) of a second LSTM unit (602), wherein the first and second trained model parameters (219) are obtained based on training of the text classification system (101) using a plurality of similar pair of texts and a plurality of dissimilar pair of texts from the plurality of texts (212);
combine the first vector representation (213) and the second vector representation (213) using a combiner operator (222) to obtain a combined vector representation (214), wherein the combiner operator (222) is selected from a plurality of combiner operators (222) based on the training using accuracy (221) of a classifier (104) of the text classification system (101); and
provide the combined vector representation (214) to the classifier (104) for determining a classification (215) of the text.

8. The text classification system (101) as claimed in claim 7 and further comprises the processor to clean the text to remove irrelevant data from the text.

9. The text classification system (101) as claimed in claim 7, wherein the training using the similar pair of texts and the dissimilar pair of texts, comprises:
receiving a first text and a second text, wherein the first text and the second text are one of the plurality of similar pair of texts and one of the plurality of dissimilar pair of texts;
generating a first labeled text (216) for the first text and a second labeled text (216) for the second text based on a plurality of labels (217) retrieved from data source 105 associated with the text classification system (101);
generating a third vector representation (213) of the first text using first model parameters (218) of the first LSTM unit (601) and a fourth vector representation (218) of the second text using second model parameters (218) of the second LSTM unit (602); and
modifying the first model parameters (218) and the second model parameters (218) to obtain the first trained model parameters (219) and the second trained model parameters (219) respectively based on norm distance (220) associated between the third vector representation (213) and the fourth vector representation (213).

10. The text classification system (101) as claimed in claim 9, wherein the first labeled text (216) comprises the first text and one of the plurality of labels (217) associated with the first text and the second labeled text (216) comprises the second text and one of the plurality of labels (217) associated with the second text.

11. The text classification system (101) as claimed in claim 9, wherein modifying the first model parameters (218) and the second model parameters (218) comprises proximating exponential of the negative of norm distance (220) to value zero for the dissimilar pair of texts and proximating the exponential of the negative of norm distance (220) to value one for the similar pair of texts.

12. The text classification system (101) as claimed in claim7, wherein selecting the combiner operator 222 comprises:
receiving each of the plurality of texts (212);
generating a fifth vector representation (213) for each of the plurality of texts (212) using the first trained model parameters (219) of the first LSTM unit (601) and a sixth vector representation (213) for each of the plurality of texts (212) using the second trained model parameters (219) of the second LSTM unit (602);
combining, by the text classification system, the third vector representation (213) and the fourth vector representation (213) using each of the plurality of combiner operators (222) to obtain corresponding a plurality of combined vector representations (214);
providing, by the text classification system, the plurality of combined vector representations (214) to the classifier (104) to obtain classification (215) for each of the plurality of combined vector representations (214);
determining, by the text classification system, accuracy (221) for each of the classification (215); and
selecting, by the text classification system, the combiner operator (222) from the plurality of combiner operators (222), based on the accuracy (221) of the classification (215) of the plurality of combined vector representations (214).

Dated this 17th day of February 2017

R Ramya Rao
Of K&S Partners
Agent for the Applicant , Description:TECHNICAL FIELD
The present subject matter is related in general to the field of determining classification, more particularly, but not exclusively to a method and system for converting text to a vector representation for determining classification of the text.

Documents

Orders

Section Controller Decision Date

Application Documents

# Name Date
1 201741005770-IntimationOfGrant18-01-2023.pdf 2023-01-18
1 Power of Attorney [17-02-2017(online)].pdf 2017-02-17
2 201741005770-PatentCertificate18-01-2023.pdf 2023-01-18
2 Form 5 [17-02-2017(online)].pdf 2017-02-17
3 Form 3 [17-02-2017(online)].pdf 2017-02-17
3 201741005770-Written submissions and relevant documents [13-12-2022(online)].pdf 2022-12-13
4 Form 18 [17-02-2017(online)].pdf_282.pdf 2017-02-17
4 201741005770-AMENDED DOCUMENTS [01-11-2022(online)].pdf 2022-11-01
5 Form 18 [17-02-2017(online)].pdf 2017-02-17
5 201741005770-Correspondence to notify the Controller [01-11-2022(online)].pdf 2022-11-01
6 Drawing [17-02-2017(online)].pdf 2017-02-17
6 201741005770-FORM 13 [01-11-2022(online)].pdf 2022-11-01
7 Description(Complete) [17-02-2017(online)].pdf_281.pdf 2017-02-17
7 201741005770-POA [01-11-2022(online)].pdf 2022-11-01
8 Description(Complete) [17-02-2017(online)].pdf 2017-02-17
8 201741005770-US(14)-HearingNotice-(HearingDate-29-11-2022).pdf 2022-10-27
9 201741005770-FER.pdf 2021-10-17
9 REQUEST FOR CERTIFIED COPY [22-02-2017(online)].pdf 2017-02-22
10 201741005770-ABSTRACT [27-07-2021(online)].pdf 2021-07-27
10 PROOF OF RIGHT [31-05-2017(online)].pdf 2017-05-31
11 201741005770-CLAIMS [27-07-2021(online)].pdf 2021-07-27
11 Correspondence by Agent_Form 1_01-06-2017.pdf 2017-06-01
12 201741005770 -Abstract.jpg 2017-06-07
12 201741005770-COMPLETE SPECIFICATION [27-07-2021(online)].pdf 2021-07-27
13 201741005770-CORRESPONDENCE [27-07-2021(online)].pdf 2021-07-27
13 201741005770-RELEVANT DOCUMENTS [27-07-2021(online)].pdf 2021-07-27
14 201741005770-DRAWING [27-07-2021(online)].pdf 2021-07-27
14 201741005770-PETITION UNDER RULE 137 [27-07-2021(online)].pdf 2021-07-27
15 201741005770-FER_SER_REPLY [27-07-2021(online)].pdf 2021-07-27
15 201741005770-OTHERS [27-07-2021(online)].pdf 2021-07-27
16 201741005770-FORM 3 [27-07-2021(online)].pdf 2021-07-27
16 201741005770-Information under section 8(2) [27-07-2021(online)].pdf 2021-07-27
17 201741005770-Information under section 8(2) [27-07-2021(online)].pdf 2021-07-27
17 201741005770-FORM 3 [27-07-2021(online)].pdf 2021-07-27
18 201741005770-FER_SER_REPLY [27-07-2021(online)].pdf 2021-07-27
18 201741005770-OTHERS [27-07-2021(online)].pdf 2021-07-27
19 201741005770-DRAWING [27-07-2021(online)].pdf 2021-07-27
19 201741005770-PETITION UNDER RULE 137 [27-07-2021(online)].pdf 2021-07-27
20 201741005770-CORRESPONDENCE [27-07-2021(online)].pdf 2021-07-27
20 201741005770-RELEVANT DOCUMENTS [27-07-2021(online)].pdf 2021-07-27
21 201741005770 -Abstract.jpg 2017-06-07
21 201741005770-COMPLETE SPECIFICATION [27-07-2021(online)].pdf 2021-07-27
22 201741005770-CLAIMS [27-07-2021(online)].pdf 2021-07-27
22 Correspondence by Agent_Form 1_01-06-2017.pdf 2017-06-01
23 201741005770-ABSTRACT [27-07-2021(online)].pdf 2021-07-27
23 PROOF OF RIGHT [31-05-2017(online)].pdf 2017-05-31
24 REQUEST FOR CERTIFIED COPY [22-02-2017(online)].pdf 2017-02-22
24 201741005770-FER.pdf 2021-10-17
25 Description(Complete) [17-02-2017(online)].pdf 2017-02-17
25 201741005770-US(14)-HearingNotice-(HearingDate-29-11-2022).pdf 2022-10-27
26 Description(Complete) [17-02-2017(online)].pdf_281.pdf 2017-02-17
26 201741005770-POA [01-11-2022(online)].pdf 2022-11-01
27 Drawing [17-02-2017(online)].pdf 2017-02-17
27 201741005770-FORM 13 [01-11-2022(online)].pdf 2022-11-01
28 Form 18 [17-02-2017(online)].pdf 2017-02-17
28 201741005770-Correspondence to notify the Controller [01-11-2022(online)].pdf 2022-11-01
29 Form 18 [17-02-2017(online)].pdf_282.pdf 2017-02-17
29 201741005770-AMENDED DOCUMENTS [01-11-2022(online)].pdf 2022-11-01
30 Form 3 [17-02-2017(online)].pdf 2017-02-17
30 201741005770-Written submissions and relevant documents [13-12-2022(online)].pdf 2022-12-13
31 201741005770-PatentCertificate18-01-2023.pdf 2023-01-18
31 Form 5 [17-02-2017(online)].pdf 2017-02-17
32 201741005770-IntimationOfGrant18-01-2023.pdf 2023-01-18
32 Power of Attorney [17-02-2017(online)].pdf 2017-02-17

Search Strategy

1 2021-01-2914-08-39E_29-01-2021.pdf

ERegister / Renewals

3rd: 17 Apr 2023

From 17/02/2019 - To 17/02/2020

4th: 17 Apr 2023

From 17/02/2020 - To 17/02/2021

5th: 17 Apr 2023

From 17/02/2021 - To 17/02/2022

6th: 17 Apr 2023

From 17/02/2022 - To 17/02/2023

7th: 17 Apr 2023

From 17/02/2023 - To 17/02/2024

8th: 15 Feb 2024

From 17/02/2024 - To 17/02/2025

9th: 15 Feb 2025

From 17/02/2025 - To 17/02/2026