Sign In to Follow Application
View All Documents & Correspondence

Method And System Of Predicting Worst Affected Cells Or Base Stations In A Network

Abstract: ABSTRACT METHOD AND SYSTEM OF PREDICTING WORST AFFECTED CELLS OR BASE STATIONS IN A NETWORK The present disclosure relates to a system (120) and a method (500) of predicting worst affected cells or base stations in a network (105). The method (500) includes the step of collecting data from base stations and aggregating the data. The method (500) includes the step of extracting features from the data. The method (500) further includes the step of splitting the data into a training dataset and a testing dataset. The method (500) further includes the step of training the training dataset using one or more machine learning models. The method (500) further includes the step of detecting unusual patterns or relationships between the features in the data. Further the method (500) includes the step of assessing the unusual patterns or relationships detected for predicting worst affected cells or the base stations in the network (105). Ref. Fig. 5

Get Free WhatsApp Updates!
Notices, Deadlines & Correspondence

Patent Information

Application #
Filing Date
07 October 2023
Publication Number
20/2025
Publication Type
INA
Invention Field
COMPUTER SCIENCE
Status
Email
Parent Application

Applicants

JIO PLATFORMS LIMITED
Office-101, Saffron, Nr. Centre Point, Panchwati 5 Rasta, Ambawadi, India

Inventors

1. Aayush Bhatnagar
Reliance Corporate Park, Thane - Belapur Road, Ghansoli, Navi Mumbai, Maharashtra 400701, India
2. Ankit Murarka
Reliance Corporate Park, Thane - Belapur Road, Ghansoli, Navi Mumbai, Maharashtra 400701, India
3. Jugal Kishore
Reliance Corporate Park, Thane - Belapur Road, Ghansoli, Navi Mumbai, Maharashtra 400701, India
4. Chandra Ganveer
Reliance Corporate Park, Thane - Belapur Road, Ghansoli, Navi Mumbai, Maharashtra 400701, India
5. Sanjana Chaudhary
Reliance Corporate Park, Thane - Belapur Road, Ghansoli, Navi Mumbai, Maharashtra 400701, India
6. Gourav Gurbani
Reliance Corporate Park, Thane - Belapur Road, Ghansoli, Navi Mumbai, Maharashtra 400701, India
7. Yogesh Kumar
Reliance Corporate Park, Thane - Belapur Road, Ghansoli, Navi Mumbai, Maharashtra 400701, India
8. Avinash Kushwaha
Reliance Corporate Park, Thane - Belapur Road, Ghansoli, Navi Mumbai, Maharashtra 400701, India
9. Dharmendra Kumar Vishwakarma
Reliance Corporate Park, Thane - Belapur Road, Ghansoli, Navi Mumbai, Maharashtra 400701, India
10. Sajal Soni
Reliance Corporate Park, Thane - Belapur Road, Ghansoli, Navi Mumbai, Maharashtra 400701, India
11. Niharika Patnam
Reliance Corporate Park, Thane - Belapur Road, Ghansoli, Navi Mumbai, Maharashtra 400701, India
12. Shubham Ingle
Reliance Corporate Park, Thane - Belapur Road, Ghansoli, Navi Mumbai, Maharashtra 400701, India
13. Harsh Poddar
Reliance Corporate Park, Thane - Belapur Road, Ghansoli, Navi Mumbai, Maharashtra 400701, India
14. Sanket Kumthekar
Reliance Corporate Park, Thane - Belapur Road, Ghansoli, Navi Mumbai, Maharashtra 400701, India
15. Mohit Bhanwria
Reliance Corporate Park, Thane - Belapur Road, Ghansoli, Navi Mumbai, Maharashtra 400701, India
16. Shashank Bhushan
Reliance Corporate Park, Thane - Belapur Road, Ghansoli, Navi Mumbai, Maharashtra 400701, India
17. Vinay Gayki
Reliance Corporate Park, Thane - Belapur Road, Ghansoli, Navi Mumbai, Maharashtra 400701, India
18. Aniket Khade
Reliance Corporate Park, Thane - Belapur Road, Ghansoli, Navi Mumbai, Maharashtra 400701, India
19. Durgesh Kumar
Reliance Corporate Park, Thane - Belapur Road, Ghansoli, Navi Mumbai, Maharashtra 400701, India
20. Zenith Kumar
Reliance Corporate Park, Thane - Belapur Road, Ghansoli, Navi Mumbai, Maharashtra 400701, India
21. Gaurav Kumar
Reliance Corporate Park, Thane - Belapur Road, Ghansoli, Navi Mumbai, Maharashtra 400701, India
22. Manasvi Rajani
Reliance Corporate Park, Thane - Belapur Road, Ghansoli, Navi Mumbai, Maharashtra 400701, India
23. Kishan Sahu
Reliance Corporate Park, Thane - Belapur Road, Ghansoli, Navi Mumbai, Maharashtra 400701, India
24. Sunil meena
Reliance Corporate Park, Thane - Belapur Road, Ghansoli, Navi Mumbai, Maharashtra 400701, India
25. Supriya Kaushik De
Reliance Corporate Park, Thane - Belapur Road, Ghansoli, Navi Mumbai, Maharashtra 400701, India
26. Kumar Debashish
Reliance Corporate Park, Thane - Belapur Road, Ghansoli, Navi Mumbai, Maharashtra 400701, India
27. Mehul Tilala
Reliance Corporate Park, Thane - Belapur Road, Ghansoli, Navi Mumbai, Maharashtra 400701, India
28. Satish Narayan
Reliance Corporate Park, Thane - Belapur Road, Ghansoli, Navi Mumbai, Maharashtra 400701, India
29. Rahul Kumar
Reliance Corporate Park, Thane - Belapur Road, Ghansoli, Navi Mumbai, Maharashtra 400701, India
30. Harshita Garg
Reliance Corporate Park, Thane - Belapur Road, Ghansoli, Navi Mumbai, Maharashtra 400701, India
31. Kunal Telgote
Reliance Corporate Park, Thane - Belapur Road, Ghansoli, Navi Mumbai, Maharashtra 400701, India
32. Ralph Lobo
Reliance Corporate Park, Thane - Belapur Road, Ghansoli, Navi Mumbai, Maharashtra 400701, India
33. Girish Dange
Reliance Corporate Park, Thane - Belapur Road, Ghansoli, Navi Mumbai, Maharashtra 400701, India

Specification

DESC:
FORM 2
THE PATENTS ACT, 1970
(39 of 1970)
&
THE PATENTS RULES, 2003

COMPLETE SPECIFICATION
(See section 10 and rule 13)
1. TITLE OF THE INVENTION
METHOD AND SYSTEM OF PREDICTING WORST AFFECTED CELLS OR BASE STATIONS IN A NETWORK
2. APPLICANT(S)
NAME NATIONALITY ADDRESS
JIO PLATFORMS LIMITED INDIAN OFFICE-101, SAFFRON, NR. CENTRE POINT, PANCHWATI 5 RASTA, AMBAWADI, AHMEDABAD 380006, GUJARAT, INDIA
3.PREAMBLE TO THE DESCRIPTION

THE FOLLOWING SPECIFICATION PARTICULARLY DESCRIBES THE NATURE OF THIS INVENTION AND THE MANNER IN WHICH IT IS TO BE PERFORMED.

FIELD OF THE INVENTION
[0001] The present invention relates to the field of network monitoring and analysis, and more particularly relates to a system and a method of predicting worst affected cells or base stations in a network.
BACKGROUND OF THE INVENTION
[0002] With the increase in number of users, the network service provisions have been implementing to up-gradations to enhance the service quality so as to keep pace with such high demand. With advancement of technology, there is a demand for the telecommunication service to induce up to date features into the scope of provision so as to enhance user experience and implement advanced monitoring mechanisms. There are regular data analyses to observe issues beforehand for which many data collection as well as assessment practices are implemented in a network.
[0003] Probing units, a crucial component within telecommunications networks, are tasked with the responsibility of data analytics, providing valuable insights into network performance, and particularly focusing on Raw Call Release Reason data. One of its key functions is to collect and analyse this raw data to pinpoint Next-Generation Node B with the highest count of released calls, which are essentially the network nodes facing the most call drops. The gNBs having independent network function is responsible for radio communication with UEs in its coverage area.
[0004] However, in traditional systems, identifying these problematic gNBs is a manual process that necessitates the analysis of the Call Release Reason (CRR) data for the gNodeB. Only after this analysis, the probing unit notifies the telecom operators to take corrective actions aimed at reducing call drops. The delay in addressing network issues may result in customer dissatisfaction and service disruptions.
[0005] There is a need for a mechanism, more specifically a system and method thereof to predict and flag specific network areas or cells susceptible to call drops due to issues like weak signal strength, network congestion, or resource constraints.

SUMMARY OF THE INVENTION
[0006] One or more embodiments of the present disclosure provide a method and system of predicting worst affected cells or base stations in a network.
[0007] In one aspect of the present invention, the method of predicting the worst affected cells or base stations in the network is disclosed. The method includes the step of collecting data from the base stations and aggregating the data. The method includes the step of extracting features from the data. The method further includes the step of splitting the data into a training dataset and a testing dataset. The method further includes the step of detecting unusual patterns or relationships between the features in the data. Further the method includes the step of assessing the unusual patterns or relationships detected for predicting the worst affected cells or the base stations in the network.
[0008] In an embodiment, the features comprising one of timestamps, call parameters, geographic coordinates, and network load metrics.
[0009] In an embodiment, the method includes pre-processing the data for data definition, data normalization, or data cleaning.
[0010] In an embodiment, the method includes annotating the data with a target variable for identifying the worst affected cells or base stations in the network.
[0011] In one embodiment, the method includes assessing performance of the one or more machine learning models using the testing dataset.
[0012] In an embodiment, the method includes identifying areas or cells at risk of the call releases due to one of a poor signal strength, congestion, and resource unavailability in the network.
[0013] In an embodiment, the one or more processor (205) trains one or more machine learning models using the training dataset.
[0014] In another aspect of the present invention, the system of predicting the worst affected cells or the base stations in the network is disclosed. The system includes a probing unit configured to collect data from the base stations and aggregate the data. The system further includes a pre-processing unit configured to extract features from the data and split the data into a training dataset and a testing dataset. The system further includes an anomaly detection unit configured to detect unusual patterns or relationships between the features in the data. The system further includes a prediction unit configured to assess the unusual patterns or relationships detected for predicting the worst affected cells or the base stations in the network.
[0015] In another aspect of the invention, a non-transitory computer-readable medium having stored thereon computer-readable instructions is disclosed. The computer-readable instructions are executed by a processor. The processor is configured to collect data from the base stations and aggregate the data. Further, the processor is configured to extract features from the data and split the data into a training dataset and a testing dataset. Further, the processor is configured to detect unusual patterns or relationships between the features in the data. Further, the processor is configured to assess the unusual patterns or relationships detected for predicting worst affected cells or base stations in the network.
[0016] Other features and aspects of this invention will be apparent from the following description and the accompanying drawings. The features and advantages described in this summary and in the following detailed description are not all-inclusive, and particularly, many additional features and advantages will be apparent to one of ordinary skill in the relevant art, in view of the drawings, specification, and claims hereof. Moreover, it should be noted that the language used in the specification has been principally selected for readability and instructional purposes and may not have been selected to delineate or circumscribe the inventive subject matter, resort to the claims being necessary to determine such inventive subject matter.

BRIEF DESCRIPTION OF THE DRAWINGS
[0017] The accompanying drawings, which are incorporated herein, and constitute a part of this disclosure, illustrate exemplary embodiments of the disclosed methods and systems in which like reference numerals refer to the same parts throughout the different drawings. Components in the drawings are not necessarily to scale, emphasis instead being placed upon clearly illustrating the principles of the present disclosure. Some drawings may indicate the components using block diagrams and may not represent the internal circuitry of each component. It will be appreciated by those skilled in the art that disclosure of such drawings includes disclosure of electrical components, electronic components or circuitry commonly used to implement such components.
[0018] FIG. 1 is an exemplary block diagram of an environment of predicting worst affected cells or base stations in a network, according to one or more embodiments of the present invention;
[0019] FIG. 2 is an exemplary block diagram of a system for predicting the worst affected cells or the base stations in the network, according to one or more embodiments of the present invention;
[0020] FIG. 3 is an exemplary block diagram of an architecture implemented in the system of the FIG. 2, according to one or more embodiments of the present invention;
[0021] FIG. 4 is a flowchart diagram of predicting the worst affected cells or the base stations in the network, according to one or more embodiments of the present invention; and
[0022] FIG. 5 is a schematic representation of a method of predicting the worst affected cells or the base stations in the network, according to one or more embodiments of the present invention.
[0023] The foregoing shall be more apparent from the following detailed description of the invention.

DETAILED DESCRIPTION OF THE INVENTION
[0024] Some embodiments of the present disclosure, illustrating all its features, will now be discussed in detail. It must also be noted that as used herein and in the appended claims, the singular forms "a", "an" and "the" include plural references unless the context clearly dictates otherwise.
[0025] Various modifications to the embodiment will be readily apparent to those skilled in the art and the generic principles herein may be applied to other embodiments. However, one of ordinary skill in the art will readily recognize that the present disclosure including the definitions listed here below are not intended to be limited to the embodiments illustrated but is to be accorded the widest scope consistent with the principles and features described herein.
[0026] A person of ordinary skill in the art will readily ascertain that the illustrated steps detailed in the figures and here below are set out to explain the exemplary embodiments shown, and it should be anticipated that ongoing technological development will change the manner in which particular functions are performed. These examples are presented herein for purposes of illustration, and not limitation. Further, the boundaries of the functional building blocks have been arbitrarily defined herein for the convenience of the description. Alternative boundaries can be defined so long as the specified functions and relationships thereof are appropriately performed. Alternatives (including equivalents, extensions, variations, deviations, etc., of those described herein) will be apparent to persons skilled in the relevant art(s) based on the teachings contained herein. Such alternatives fall within the scope and spirit of the disclosed embodiments.
[0027] FIG. 1 illustrates an exemplary block diagram of an environment 100 of predicting worst affected cells or base stations in a network 105, according to one or more embodiments of the present disclosure. In this regard, the environment 100 includes a User Equipment (UE) 110, a server 115, the network 105 and a system 120 communicably coupled to each of monitoring the network 105. The cell refers to a geographic area served by a base station. Each cell allows the UE 110 to connect to the network 105 and access services. The network 105 is divided into multiple cells to ensure coverage over a wide area. The cells are crucial for managing user traffic efficiently, optimizing coverage, and ensuring seamless handovers as users move from one cell to another. The base stations are responsible for managing communication between the UE 110 and the core network. The base stations are part of the Radio Access Network (RAN) in the network 105 and plays the key role in handling wireless communication includes, but not limited to, data transmission, signal processing, and control functions.
[0028] The worst affected cells refer to the cells or the base stations experiencing performance degradation or ability to serve users effectively. The cells facing challenges like network congestion, interference, hardware failure, and high user density are termed as the worst affected cells.
[0029] As per the illustrated embodiment and for the purpose of description and illustration, the UE 110 includes, but not limited to, a first UE 110a, a second UE 110b, and a third UE 110c, and should nowhere be construed as limiting the scope of the present disclosure. In alternate embodiments, the UE 110 may include a plurality of UEs as per the requirement. For ease of reference, each of the first UE 110a, the second UE 110b, and the third UE 110c, will hereinafter be collectively and individually referred to as the “User Equipment (UE) 110”.
[0030] In an embodiment, the UE 110 is one of, but not limited to, any electrical, electronic, electro-mechanical or an equipment and a combination of one or more of the above devices such as a smartphone, virtual reality (VR) devices, augmented reality (AR) devices, laptop, a general-purpose computer, desktop, personal digital assistant, tablet computer, mainframe computer, or any other computing device.
[0031] The environment 100 includes the server 115 accessible via the network 105. The server 115 may include, by way of example but not limitation, one or more of a standalone server, a server blade, a server rack, a bank of servers, a server farm, hardware supporting a part of a cloud service or system, a home server, hardware running a virtualized server, one or more processors executing code to function as a server, one or more machines performing server-side functionality as described herein, at least a portion of any of the above, some combination thereof. In an embodiment, the entity may include, but is not limited to, a vendor, a network operator, a company, an organization, a university, a lab facility, a business enterprise side, a defense facility side, or any other facility that provides service.
[0032] The network 105 includes, by way of example but not limitation, one or more of a wireless network, a wired network, an internet, an intranet, a public network, a private network, a packet-switched network, a circuit-switched network, an ad hoc network, an infrastructure network, a Public-Switched Telephone Network (PSTN), a cable network, a cellular network, a satellite network, a fiber optic network, or some combination thereof. The network 105 may include, but is not limited to, a Third Generation (3G), a Fourth Generation (4G), a Fifth Generation (5G), a Sixth Generation (6G), a New Radio (NR), a Narrow Band Internet of Things (NB-IoT), an Open Radio Access Network (O-RAN), and the like.
[0033] The network 105 may also include, by way of example but not limitation, at least a portion of one or more networks 105 having one or more nodes that transmit, receive, forward, generate, buffer, store, route, switch, process, or a combination thereof, etc. one or more messages, packets, signals, waves, voltage or current levels, some combination thereof, or so forth. The network 105 may also include, by way of example but not limitation, one or more of a wireless network, a wired network, an internet, an intranet, a public network, a private network, a packet-switched network, a circuit-switched network, an ad hoc network, an infrastructure network, a Public-Switched Telephone Network (PSTN), a cable network, a cellular network, a satellite network, a fiber optic network, a VOIP or some combination thereof.
[0034] The environment 100 further includes the system 120 communicably coupled to the server 115 and the UE 110 via the network 105. The system 120 is configured to manage CNFs in the network 105. As per one or more embodiments, the system 120 is adapted to be embedded within the server 115 or embedded as an individual entity.
[0035] Operational and construction features of the system 120 will be explained in detail with respect to the following figures.
[0036] FIG. 2 is an exemplary block diagram of the system 120 of predicting the worst affected cells or the base stations in the network 105, according to one or more embodiments of the present invention.
[0037] As per the illustrated embodiment, the system 120 includes one or more processors 205, a memory 210, a user interface 215, and a database 220. For the purpose of description and explanation, the description will be explained with respect to one processor 205 and should nowhere be construed as limiting the scope of the present disclosure. In alternate embodiments, the system 120 may include more than one processor 205 as per the requirement of the network 105. The one or more processors 205, hereinafter referred to as the processor 205 may be implemented as one or more microprocessors, microcomputers, microcontrollers, digital signal processors, central processing units, state machines, logic circuitries, single board computers, and/or any devices that manipulate signals based on operational instructions.
[0038] As per the illustrated embodiment, the processor 205 is configured to fetch and execute computer-readable instructions stored in the memory 210. The memory 210 may be configured to store one or more computer-readable instructions or routines in a non-transitory computer-readable storage medium, which may be fetched and executed to create or share data packets over a network service. The memory 210 may include any non-transitory storage device including, for example, volatile memory such as RAM, or non-volatile memory such as disk memory, EPROMs, FLASH memory, unalterable memory, and the like.
[0039] In an embodiment, the user interface 215 includes a variety of interfaces, for example, interfaces for a graphical user interface, a web user interface, a Command Line Interface (CLI), and the like. The user interface 215 facilitates communication of the system 120. In one embodiment, the user interface 215 provides a communication pathway for one or more components of the system 120. Examples of such components include, but are not limited to, the UE 110 and the database 220.
[0040] The database 220 is one of, but not limited to, a centralized database, a cloud-based database, a commercial database, an open-source database, a distributed database, an end-user database, a graphical database, a No-Structured Query Language (NoSQL) database, an object-oriented database, a personal database, an in-memory database, a document-based database, a time series database, a wide column database, a key value database, a search database, a cache databases, and so forth. The foregoing examples of database 220 types are non-limiting and may not be mutually exclusive e.g., a database can be both commercial and cloud-based, or both relational and open-source, etc.
[0041] In order for the system 120 to predict the worst affected cells in the network 105, the processor 205 includes one or more modules. In one embodiment, the one or more modules/units includes, but not limited to, a probing unit 225, a pre-processing unit 230, a model training unit 235, an anomaly detection unit 240, and a prediction unit 245 communicably coupled to each other of monitoring the network 105.
[0042] In one embodiment, the one or more modules may be used in combination or interchangeably for predicting the worst affected cells or the base stations in the network 105.
[0043] The probing unit 225, the pre-processing unit 230, the model training unit 235, the anomaly detection unit 240, and the prediction unit 245, in an embodiment, are implemented as a combination of hardware and programming (for example, programmable instructions) to implement one or more functionalities of the processor 205. In the examples described herein, such combinations of hardware and programming are implemented in several different ways. For example, the programming for the processor 205 may be processor-executable instructions stored on a non-transitory machine-readable storage medium and the hardware for the processor may comprise a processing resource (for example, one or more processors), to execute such instructions. In the present examples, the memory 210 may store instructions that, when executed by the processing resource, implement the processor. In such examples, the system 120 may comprise the memory 210 storing the instructions and the processing resource to execute the instructions, or the memory 210 may be separate but accessible to the system 120 and the processing resource. In other examples, the processor 205 may be implemented by electronic circuitry.
[0044] In an embodiment, the probing unit 225 is configured to collect data from the base stations and aggregate the data. Hereinafter, the base stations are referred as Next-Generation Node B (gNBs). Further the base stations or the gNBs can be used interchangeably without limiting the scope of the invention. In one embodiment, the base stations are configured to continuously transmit the data to the probing unit 225 on periodic basis. For example, the probing unit 225 receives the data from the one base stations in every one hour. In yet another embodiment, the probing unit 225 collects the data from the base stations via an interface. In one embodiment, the interface includes at least one of, but not limited to, one or more Application Programming Interfaces (APIs) which are used by the probing unit 225 for collecting the data. The one or more APIs are sets of rules and protocols that allow different entities to communicate with each other. The one or more APIs define the methods and data formats that entities can use to request and exchange information, enabling integration and functionality across various platforms.
[0045] The data includes, but is not limited to, signal strength metrics, traffic data, network load metrics, mobility patterns, and Quality of Service metrics (QoS). The signal strength metrics are measurements of received signal strength indicator and reference signal received power to assess network coverage and quality. The traffic data includes information on data usage patterns such as, but not limited to, download and upload speeds, the volume of data transmitted, and the types of applications being used. the network load metrics encompass data on the number of active users, resource utilization, and load balancing across the gNBs. The mobility patterns include information on user movement and handover events as devices transition between different the gNBs or cells. The QoS metrics consist of data on latency, jitter, packet loss, and throughput, which are used to evaluate the performance of different services.
[0046] Upon collecting and aggregating the data from the gNBs, the pre-processing unit 230 is configured to extract features from the data. In one embodiment, the pre-processing unit 230 extract features from the data based on a predefined criteria such as the extracted features must include at least one of, but not limited to, the call parameters, the geographic coordinates, the network load metrics and timestamps. Herein, the predefined criteria is defined by at least one of, but not limited to, the user and the model based on the previous model training experience. The pre-processing unit is further configured to split the data into a training dataset and a testing dataset. The training dataset is a subset of the collected data used to train one or more machine learning model. The training dataset consists of input features and the corresponding labels, allowing the model to learn patterns and relationships within the data. The examples of the training dataset include, but not limited to, signal strength metrics and traffic patterns. The training dataset applied to the model to identify patterns and relationships in the data and adjusts the parameters of the one or more machine learning model to minimize prediction errors based on the features and labels in the training dataset. The parameters refer to the internal configurations within the one or more machine leaning models that are adjusted during the training process to improve the performance.
[0047] In an embodiment, the features include one of timestamps, call parameters, geographic coordinates, and network load metrics. The call parameters are specific metrics related to voice or video calls in the network 105. The call parameters include, but are not limited to, all setup time, call drop rate, and voice quality metrics. The geographic coordinates are specific latitude and longitude values that pinpoint the physical location of the user or the network node. the geographic coordinates include, but not limited to, location-based services and network optimization. The network load metrics are measurements that provide insight into the utilization of the network resources. The network load metrics may include, but not limited to, active users, resources utilization, data throughput, and latency.
[0048] The testing dataset is a separate subset of the collected data used to evaluate the performance of the trained model. The testing dataset helps assess how well the model generalizes to new, unseen data. The examples of testing dataset include, but not limited to, new user movement data and new QoS measurements. The model predicts outcomes based on the features in the testing dataset, and the results are compared to the actual labels to measure accuracy, precision, recall, and other performance metrics.
[0049] Subsequent to extracting the features, the pre-processing unit 230 is configured to pre-process the data collected from the base stations. The pre-processing includes at least one of, but not limited to, data definition, data normalization, or data cleaning. The data definition involves specifying the structure, format, and meaning of the data collected from the network 105. The data definition includes identifying the types of data being collected such as, but not limited to, signal strength, user location, call parameters. The data normalization is the process of transforming data into a consistent format, ensuring that the data is comparable and may be analyzed effectively. The data normalization may involve adjusting data values to a common scale without distorting differences in the ranges of values. The data cleaning refers to the process of identifying and correcting errors, inconsistencies, and inaccuracies in the data. The data normalization is critical in the network 105, where the data from various sources may be noisy or incomplete. The data normalization includes, but not limited to, removing duplicate entries, handling missing values, and correcting errors.
[0050] In one embodiment, the pre-processing unit 230 is configured to preprocess the data to ensure the data consistency and quality within the system 120. The main goal of, the pre-processing unit 230 is to achieve a standardized data format of the data with no errors. The pre-processing unit 230 eliminates duplicate data and inconsistencies which reduces manual efforts. The pre-processing unit 230 ensures that the pre-processed data is stored appropriately in at least one of, the database 220 and the pre-processed data is ready for subsequent retrieval and analysis.
[0051] Furthermore, the pre-processing unit 230 annotates the data with a target variable for identifying the worst affected cells or the gNBs in the network 105. In one embodiment, the target variables are one or more parameters related to the cells or the gNBs which includes, at least one of, but not limited to, a signal strength, a throughput, and a dropped call rate. The data annotation involves adding labels or tags to data points, which helps in identifying specific features or outcomes in the dataset. In this regard, the pre-processing unit 230 annotates the data with the target variable to highlight which cells or the gNBs are the worst affected in the network 105. For example, the target variable such as the signal strength indicates the strength of the signal in the cell. If the strength of the signal is less as compared to the predefined threshold of the signal strength, then the cell or the gNBs is identified as the worst affected cells or the gNBs. By annotating the data, the one or more machine learning models may be trained to predict network 105 performance issues, allowing them to recognize patterns associated with worst affected the gNBs based on the historical data and enabling proactive measures to mitigate issues before escalating, ultimately supporting operational efficiency by allowing network operators to prioritize resources and interventions for the cells or the gNBs that require the most attention.
[0052] Thereafter, the model training unit 235 configured to train the training dataset using the one or more machine learning models. The model training unit 235 is configured to enable the one or more machine learning models to learn patterns and relationships that can predict outcomes related to network performance. The model training unit 235 is responsible for training the one or more machine learning algorithms using the dataset. The model training unit 235 prepares the training dataset by selecting relevant features and preprocesses the data through normalization to ensure consistent input for training Features and labels are fed into the model, which learns by making predictions, calculating loss, and adjusting parameters through backpropagation and optimization algorithms like gradient descent. The process of training the dataset is repeated over multiple epochs, sometimes using batch training for efficiency, while validation sets monitor performance to prevent overfitting. After evaluating the model's accuracy and precision on the testing dataset, the trained model is deployed in the network management system to enable real-time predictions and proactive management of the network 105.
[0053] For example, let us consider that the training data is provided to the one or more models by the model training unit 235. Further, the one or more models generates outputs based on the provided training data. Furthermore, based on the generated outputs, the anomaly detection unit 240 assesses performance of the one or more machine learning models. Herein, the assessing refers to the process of evaluating, measuring, or estimating the quality, value, or performance of the one or more machine learning models. Based on the assessing, the model with the superior performance as compared to the other models is selected by the anomaly detection unit 240 to detect the anomaly.
[0054] Upon training the one or more machine learning models using the dataset from the model training unit 235, the anomaly detection unit 240 is configured to detect unusual patterns or relationships between the historical data or real-time data. The anomaly detection unit 240 identifies unusual patterns in the network data by collecting metrics from sources like the gNBs and user devices, including signal strength and latency. The networks data generates the information and metrics within the network environment. The anomaly detection unit 240 extracts the feature and establishes a standard for normal behavior using historical data. The anomaly detection unit 240 then applies algorithms such as, but not limited to, statistical methods, isolation forests, and clustering techniques to detect deviations in real-time. Upon detecting an anomaly, the anomaly detection unit 240 generates alerts for network operators, detailing the issue and affected the gNBs. The detection process involves collecting metrics and data from various sources, extracting relevant features, and establishing the standard for normal behavior using the historical data. Algorithms such as, but not limited to, statistical methods, isolation forests, and clustering techniques are then applied to analyze real-time data against the standard for normal behaviour, identifying significant deviations that may indicate anomalies. Furthermore, a feedback mechanism allows operators to validate the anomalies, refining the detection algorithms and improving future accuracy to maintain network performance and reliability.
[0055] For example, let us consider that the features in the data includes at least one of, but not limited to, call parameters, and network load metrics. Further, the call parameters may include at least one of, but not limited to, a call drop and the network load metric include at least one of, but not limited to, a latency. Herein, the relationships between two or more features, shows how changes in one metric are linked to changes in another. More particularly, the relationships due to the changes in the latency are linked to the call drop. In order to detect, the unusual patterns or relationships between the features in the data, the anomaly detection unit 240 compares the features with the standard for normal pattern or behavior of the features. Based on comparison, the anomaly detection unit 240 detects the unusual patterns or relationships between the features in the data such as even though the latency is less, the number of the call drops are more in the particular cell.
[0056] In an embodiment, the anomaly detection unit 240 assesses performance of the one or more machine learning models using the testing dataset. The anomaly detection unit 240 assesses the performance of the one or more machine learning models using the testing dataset, which consists of data that the models have not encountered during training. By applying the trained models to the testing dataset, the anomaly detection unit 240 makes predictions about whether data points are normal or anomalous. The anomaly detection unit 240 compares the predictions to the actual labels, which indicate the true classifications. Various performance metrics are calculated to quantify the models' effectiveness including accuracy, precision, recall, providing insight into how well the models generalize to new data. The process of evaluating the performance metrics of the anomaly detection unit 240 which helps to identify strengths and weaknesses, allowing for adjustments to improve accuracy such as, bit not limited to, refining feature selection or retraining with additional data. The example of the strength includes, but not limited to, real-time monitoring, predictive maintenance, enhanced user experience, and automated decision making. The examples of the weaknesses include, but not limited to, false positives, dependency on historical data, and complexity of implementation. Based on the evaluation results, the anomaly detection unit 240 select the best-performing model for deployment or tune hyperparameters to enhance performance, ensuring that the models reliably detect unusual patterns in the network 105 and contribute to effective network management and operational efficiency.
[0057] Upon detecting the unusual patterns or relationships between the features in the data, the prediction unit 245 is configured to assess the unusual patterns or relationships detected for predicting the worst affected cells or the gNBs in the network 105 that are likely to experience call releases.
[0058] The call releases occur when the active call is terminated prematurely due to various reasons, including poor network conditions, inadequate signal strength, or resource unavailability in the network 105. Frequent call releases may lead to the negative user experience and dissatisfaction among customers. By identifying specific cells or gNBs at the higher risk of causing call releases, the prediction unit 245 enables proactive measures to be taken by network operators.
[0059] The prediction unit 245 assesses unusual patterns or relationships in network data to predict which cells or the gNBs are likely to experience call releases. The prediction unit 245 analyzes various metrics such as, but not limited to, signal strength, user mobility patterns, and traffic load metrics to identify correlations with past call drops. By employing the one or more machine learning algorithms, the prediction unit 245 recognizes patterns that indicate potential connectivity issues, such as low signal strength combined with high traffic. Continuous real-time monitoring allows the prediction unit 245 to apply predictive models trained on historical data to forecast at-risk the gNBs. The prediction unit 245 generating alerts for network operators when unusual conditions arise, like sudden drops in signal quality or congestion during peak hours. For example, if the gNBs in the densely populated area is nearing capacity during the major event, the prediction unit 245 can predict call releases, enabling operators to take proactive measures like reallocating resources or optimizing load balancing. The capability of the prediction unit 245 helps maintain service quality and enhances user experience within the network 105.
[0060] In yet another example, let us consider that there are 1 lakh users connected with the base station the network 105 and the threshold value for handling the capacity of the users by the base station is 2 lakhs users. Herein, the threshold is set based on assessing the unusual patterns or relationships between the features in the data. In one scenario at a peak time the number of the number of the users connected with the base station are increasing. So, the prediction unit 245 identifies the pattern of the increasing number of the users with respect to time using the one or more models. Based on the identified pattern, the prediction unit 245 predicts that in some time interval the signal strength of the base station will be degraded due to the high traffic of the user which will lead to various call drops. So, prediction unit 245 infers the base station as the worst affected cell or the base station.
[0061] Subsequently, the prediction unit 245 identifies areas or cells at risk of the call releases due to one of a poor signal strength, congestion, and resource unavailability in the network 105. The prediction unit 245 monitors signal strength metrics collected from user devices and the gNBs. If the signal strength falls below the predefined threshold, then the prediction unit 245 identifies areas or cells at risk of the call releases due to at least one of the poor signal strengths and indicates that users may struggle to maintain the stable connection, leading to potential call drops. The predefined threshold represents the minimum signal strength required for reliable communication in the network 105 when the predefined threshold is exceeded, users may expect the high-quality connection. For example, if the cell shows a consistent drop in signal quality during peak usage times, the prediction unit 245 may be flagged as at risk. The prediction unit 245 assesses traffic load metrics, which include the number of active users and the volume of data being transmitted. The congestion refers to the situation where the demand for network resources exceeds the available capacity, leading to a degradation of service quality. The congestion results in delays, dropped calls, or failures in data transmission, as the struggles to accommodate the number of active users and the data requirements. The example of congestion includes, but not limited to, mobile network congestion, wi-fi network congestion, data center congestion, and transport network congestion. Further, the prediction unit 245 evaluates the availability of network resources, such as bandwidth and processing power. When the availability of network resources is insufficient to handle the current load, calls may be released because the network cannot maintain the required quality of service. Overloading of the gNBs or temporary unavailability of certain resources due to the maintenance or other issues can lead to such situations.
[0062] In one embodiment, upon prediction, the one or more models is retrained based on updated data received from the base stations due to which the one or more models are getting advances and leads to increase in the accuracy of the prediction. In particular, the updated data is collected from the base stations on periodic basis. The updated data is fed to the one or more models by the model training unit 235, so that based on the updated data, the model is trained again.
[0063] FIG. 3 is an exemplary block diagram of an architecture implemented in the system of the FIG. 2, according to one or more embodiments of the present invention;
[0064] The architecture 300 includes a probing unit 305, a data consumer 310, a processing hub 315, a data integration 320, a data pre-processing 325, a model training 330, a prediction module 335, a processing hub interface 340, and the database 220.
[0065] The process begins with the probing unit 305, which acts as the data collection component, gathering information from gNodeBs. The probing unit 305 data includes key features such as, but not limited to, timestamps, call parameters, geographic coordinates, and network load metrics. After the data collection, the data is passed to the data integration 320, where the collected data is aggregated and organized for further analysis. The data aggregation and organization process performed by the data integration 320 and consolidates the data collected from the gNBs preparing for further analysis like feature extraction and model training. Extraction of features is essential for understanding patterns in the network 105, such as, but not limited to, identifying potential issues with signal strength or network congestion.
[0066] Once the data is received from the probing unit 305, the data integration 320 undergoes aggregation, where data from the gNBs is combined and aligned correctly. The aggregation process may include, but not limited to, merging data, performing initial consistency checks, and preparing for the pre-processing. The data integration 320 ensures that the data is consistent and ready for the one or more machine learning operations, where the data collected from the gNBs is aggregated for subsequent analysis.
[0067] Upon data integration, the data moves to the data pre-processing 325, where the data pre-processing 325 performs essential tasks such as, but not limited to, data definition, data normalization, and data cleaning are performed. The data definition ensures that the data filed are clearly defined and in the proper format. The data normalization involves scaling the data to ensure that all features are on a consistent scale, which improves the performance of the one or more machine learning models. The data cleaning and performed involves removing or handling missing data and outliers to ensure the quality of the database 220. The present step preparing the database 220 for model training by ensuring that is clean, normalized, and ready for use.
[0068] Thereafter, the data is split into the training dataset and the testing dataset. The data ensures that the one or more machine learning models are trained on one portion of the data while another portion is reserved for evaluating the performance of the models. The training dataset is used to build the predictive model, and the testing dataset is used to evaluate the accuracy of the predictions.
[0069] Upon splitting the data, the model training 330 uses the one or more machine learning models to train on the training dataset. Specifically, the model training 330 involves analyzing the data to extract patterns, trends, and relationships from the features, such as, but not limited to, call parameters, geographic coordinates, and network load metrics. During the model training 330 phase, the one or more machine learning model learns from the data to identify underlying patterns that may later be used to make predictions or detect anomalies. The choice of the one or more machine learning algorithms applied at the stage depends on the complexity and volume of the data.
[0070] After the model is trained, the prediction module 335 is responsible for detecting unusual patterns or relationships between the features in the data. The prediction module 335 identifies potential anomalies or deviations from normal behavior such as, but not limited to, increased call drops, poor signal strength, congestion, or resource unavailability. Based on the detections, the gNodeBs or cells at the highest risk of experiencing issues such as, call releases, are predicted.
[0071] The prediction module 335 is then tested using the testing dataset to assess its accuracy. The prediction module 335 evaluates how well the model performs at identifying affected cells and the gNBs by comparing predictions with actual data. The various metrics such as, but not limited to, precision, recall, accuracy, and the root mean square error to assess the model’s effectiveness.
[0072] Both the training results and predictions generated by the one or more machine learning models are stored in the database 220. The database 220 holds historical data and can be used to re-train models over time or for further analysis. The database 220 ensures that there is the persistent record of the predictions, which can be accessed for future comparisons or investigations.
[0073] Thereafter, the anomaly detected results are sending and predictions to the data consumers 310. The data consumers 310 could be network operators or automation systems that monitor network performance and take proactive actions based on the predictions. For example, the data consumers may use this information to adjust network resources, troubleshoot poor signal areas, or mitigate congestion before leads to significant service degradation.
[0074] The processing hub interface 340 serves as the central communication link between various components within the architecture 300. The processing hub interface 340 role is to manage the flow of the data between modules such as the data integration 320, the data pre-processing 325, the model training 330, and the prediction module 335. The processing hub interface 340 ensures that data is transferred efficiently between the processing hub 315, enabling seamless interaction.
[0075] FIG. 4 is an exemplary flowchart diagram of predicting the worst affected cells or the gNBs in the network 105, according to one or more embodiments of the present invention.
[0076] At step 405, the process begins with the collection and aggregation of the raw data from the gNBs in the network 105. The data includes, but is not limited to, call release reasons, timestamps, call parameters, network load metrics, and other relevant information. The purpose of this step is to gather essential information that will later be used for training and prediction.
[0077] At step 410, upon collecting and aggregation of the data, the raw data is processed to extract key features such as, but not limited to, timestamps, geographic coordinates, call parameters, and network load metrics. The data is then divided into the training dataset and the testing dataset. This step ensures the creation of meaningful input features for the one or more machine learning models and prepares the data for training and testing.
[0078] At step 415, once the data is processed to extract the features, the training dataset is used to train the one or more machine learning models. The prediction module 335 identifies the patterns, relationships, and connections between variables in the data which includes, but not limited to, factors contributing to call releases or network congestion. The objective is to develop the prediction module 335 capable of recognizing patterns linked to potential network issues, thereby enabling early detection and response to problems.
[0079] At step 420, upon training the model, the model training 330 is evaluated against the testing dataset to assess the accuracy and performance. Various metrics, such as precision, recall, and root mean square error, are used to determine how effectively the model can predict the worst-affected cells or gNodeBs. The goal is to ensure the model's capability to accurately identify potential network issues and forecast which cells or the gNBs are most at risk.
[0080] At step 425, upon evaluating against the testing dataset to assess the accuracy and performance, the trained model produces insights by identifying cells or the gNBs that are at the highest risk of facing issues, such as call releases. These insights are based on the unusual patterns or anomalies detected during the training and testing phases. The objective is to predict the worst-affected areas in the network 105 and highlight potential problems before they result in disruptions.
[0081] At step 430, upon predicting the wort affected cell or the gNbs in the network 105, the one or more machine learning model continuously adapts based on new data and changing network conditions to ensure the one or more machine learning model remains effective in predicting network issues as conditions evolve over time.
[0082] FIG. 5 is a schematic representation of a method 500 of predicting worst affected cells or the gNBs in the network 105, according to one or more embodiments of the present invention. For the purpose of description, the method 500 is described with the embodiments as illustrated in FIG. 2 and should nowhere be construed as limiting the scope of the present disclosure.
[0083] At step 505, the method 500 includes the step of collecting data from the gNBs and aggregating the data. The data includes, but not limited to, signal strength metrics, traffic data, network load metrics, mobility patterns, and QoS.
[0084] At step 510, the method 500 includes the step of extracting features from the data. The features include one of the timestamps, call parameters, geographic coordinates, and network load metrics.
[0085] The step 515, the method 500 includes the step of splitting the data into the training dataset and the testing dataset. The training dataset is the subset of the collected data used to train the one or more machine learning model. The training dataset applied to the model to identify patterns and relationships in the data and adjusts the one or more machine learning model parameters to minimize prediction errors based on the features and labels in the training dataset. The testing dataset is a separate subset of the collected data used to evaluate the performance of the trained model. The model predicts outcomes based on the features in the testing dataset, and the results are compared to the actual labels to measure accuracy, precision, recall, and other performance metrics.
[0086] At step 520, the method 500 includes the step of training the one or more machine learning models using dataset. The process of training the dataset is repeated over multiple epochs, sometimes using batch training for efficiency, while validation sets monitor performance to prevent overfitting. After evaluating the model's accuracy and precision on the testing dataset, the trained model is deployed in the network management system to enable real-time predictions and proactive management of the network 105.
[0087] At step 525, the method 500 includes the step of detecting unusual patterns or relationships between the features in the data. By collecting metrics from sources like the gNBs and user devices, including signal strength and latency. The anomaly detection unit 240 extracts feature and establishes the normal behavior baseline using historical data, then applies algorithms such as, but not limited to, statistical methods, isolation forests, and clustering techniques to detect deviations in real-time. Upon detecting an anomaly, the anomaly detection unit 240 unit generates alerts for network operators, detailing the issue and affected the gNBs. Further, the feedback loop allows operators to validate anomalies, refining detection algorithms and improving future accuracy to maintain network performance and reliability.
[0088] At step 530, the method 500 includes the step of assessing the unusual patterns or relationships detected for predicting worst affected cells or the gNBs in the network that are likely to experience call releases. The prediction unit 245 assesses unusual patterns or relationships in network data to predict which cells or the gNBs are likely to experience call releases. The prediction unit 245 analyzes various metrics such as, but not limited to, signal strength, user mobility patterns, and traffic load metrics to identify correlations with past call drops. By employing the one or more machine learning algorithms, the prediction unit 245 recognizes patterns that indicate potential connectivity issues, such as low signal strength combined with high traffic. Continuous real-time monitoring allows the prediction unit 245 to apply predictive models trained on historical data to forecast at-risk the gNBs. The prediction unit 245 generating alerts for network operators when unusual conditions arise, like sudden drops in signal quality or congestion during peak hours.

[0089] The present invention further discloses a non-transitory computer-readable medium having stored thereon computer-readable instructions. The computer-readable instructions are executed by the processor 205. The processor 205 is configured to collect data from the base stations and aggregate the data. Further, the processor 205 is configured to extract features from the data and split the data into the training dataset and the testing dataset. Further, the processor 205 is configured to detect unusual patterns or relationships between the features in the data. Further, the processor 205 is configured to assess the unusual patterns or relationships detected for predicting worst affected cells or the base stations in the network 105 that are likely to experience call releases.
[0090] A person of ordinary skill in the art will readily ascertain that the illustrated embodiments and steps in description and drawings (FIG.1-5) are set out to explain the exemplary embodiments shown, and it should be anticipated that ongoing technological development will change the manner in which particular functions are performed. These examples are presented herein for purposes of illustration, and not limitation. Further, the boundaries of the functional building blocks have been arbitrarily defined herein for the convenience of the description. Alternative boundaries can be defined so long as the specified functions and relationships thereof are appropriately performed. Alternatives (including equivalents, extensions, variations, deviations, etc., of those described herein) will be apparent to persons skilled in the relevant art(s) based on the teachings contained herein. Such alternatives fall within the scope and spirit of the disclosed embodiments.
[0091] The present disclosure includes technical advancement by improving predictive accuracy with advanced the machine learning models and diverse data sources, such as call parameters and geographic coordinates. Automated anomaly detection and targeted risk identification address potential issues like poor signal strength and congestion. The invention robust the data pre-processing ensures high-quality inputs for reliable predictions, while systematic performance assessments facilitate continuous improvement. The scalable architecture optimizes resource allocation, reducing costs and boosting user satisfaction, and computer-readable instructions allow for easy integration across platforms, promoting widespread adoption and overall efficiency.
[0092] The present invention offers multiple advantages by enabling proactive issue resolution, allowing operators to identify and address potential problems like poor signal strength or congestion before they escalate. The results in improved network quality, fewer call drops, and higher customer satisfaction. By anticipating issues, the invention significantly reduces service disruptions, leading to more reliable services. It also optimizes resource utilization, ensuring efficient allocation of bandwidth and capacity, ultimately driving cost savings and fostering greater customer loyalty.
[0093] The present invention offers multiple advantages over the prior art and the above listed are a few examples to emphasize on some of the advantageous features. The listed advantages are to be read in a non-limiting manner.

REFERENCE NUMERALS

[0094] Environment- 100
[0095] User Equipment (UE)- 110
[0096] Server- 115
[0097] Network- 105
[0098] System -120
[0099] Processor- 205
[00100] Memory- 210
[00101] User interface- 215
[00102] Database – 220
[00103] Probing unit - 225
[00104] Pre-processing unit - 230
[00105] Model training unit - 235
[00106] Anomaly detection unit - 240
[00107] Prediction unit – 245
[00108] Probing unit – 305
[00109] Data consumers - 310
[00110] Processing hub - 315
[00111] Data integration - 320
[00112] Data pre-processing - 325
[00113] Model training - 330
[00114] Prediction module - 335
[00115] Processing hub interface - 340

,CLAIMS:CLAIMS
We Claim:
1. A method (500) of predicting worst affected cells or base stations in a network (105), the method (500) comprising the steps of:
collecting, by a one or more processor (205), data from the base stations and aggregating the data;
extracting, by the one or more processor (205), features from the data;
splitting, by the one or more processor (205), the data into a training dataset and a testing dataset;
detecting, by the one or more processor (205), unusual patterns or relationships between the features in the data; and
assessing, by the one or more processor (205), the unusual patterns or relationships detected for predicting the worst affected cells or the base stations in the network (105).

2. The method (500) as claimed in claim 1, the features comprising one of timestamps, call parameters, geographic coordinates, and network load metrics.

3. The method (500) as claimed in claim 1, comprising pre-processing, by the one or more processor (205), the data for data definition, data normalization, or data cleaning.

4. The method (500) as claimed in claim 1, comprising annotating, by the one or more processor (205), the data with a target variable for identifying the worst affected cells or the base stations in the network (105).

5. The method (500) as claimed in claim 1, comprising assessing, by the one or more processor (205), performance of the one or more machine learning models using the testing dataset.

6. The method (500) as claimed in claim 1, comprising, identifying, by the one or more processor (205), areas or cells at risk of the call releases due to one of a poor signal strength, congestion, and resource unavailability in the network (105).

7. The method (500) as claimed in claim 1, wherein the one or more processor (205) trains one or more machine learning models using the training dataset.

8. A system (120) for predicting worst affected cells or base stations in a network (105), the system (120) comprising:
a probing unit (225) configured to collect data from base stations and aggregate the data;
a pre-processing unit (230) configured to extract features from the data, and split the data into a training dataset and a testing dataset;
an anomaly detection unit (240) configured to detect unusual patterns or relationships between the features in the data; and
a prediction unit (245) configured to assess the unusual patterns or relationships detected for predicting worst affected cells or the base stations in the network (105).

9. The system (120) as claimed in claim 8, wherein the features comprise one of timestamps, call parameters, geographic coordinates, and network load metrics.

10. The system (120) as claimed in claim 8, wherein the pre-processing unit (230) pre-processes the data for data definition, data normalization, or data cleaning.

11. The system (120) as claimed in claim 8, wherein the pre-processing unit (230) annotates the data with a target variable for identifying the worst affected cells or the base stations in the network (105).

12. The system (120) as claimed in claim 8, wherein the anomaly detection unit (240) assesses performance of the one or more machine learning models using a testing dataset.

13. The system (120) as claimed in claim 8, wherein the prediction unit (245) identifies areas or cells at risk of the call releases due to one of a poor signal strength, congestion, and resource unavailability in the network (105).

14. The system (120) as claimed in claim 8, wherein a model training unit (235) is configured to train one or more machine learning models using the training dataset.

Documents

Application Documents

# Name Date
1 202321067386-STATEMENT OF UNDERTAKING (FORM 3) [07-10-2023(online)].pdf 2023-10-07
2 202321067386-PROVISIONAL SPECIFICATION [07-10-2023(online)].pdf 2023-10-07
3 202321067386-POWER OF AUTHORITY [07-10-2023(online)].pdf 2023-10-07
4 202321067386-FORM 1 [07-10-2023(online)].pdf 2023-10-07
5 202321067386-FIGURE OF ABSTRACT [07-10-2023(online)].pdf 2023-10-07
6 202321067386-DRAWINGS [07-10-2023(online)].pdf 2023-10-07
7 202321067386-DECLARATION OF INVENTORSHIP (FORM 5) [07-10-2023(online)].pdf 2023-10-07
8 202321067386-FORM-26 [27-11-2023(online)].pdf 2023-11-27
9 202321067386-Proof of Right [12-02-2024(online)].pdf 2024-02-12
10 202321067386-DRAWING [07-10-2024(online)].pdf 2024-10-07
11 202321067386-COMPLETE SPECIFICATION [07-10-2024(online)].pdf 2024-10-07
12 Abstract.jpg 2024-12-30
13 202321067386-Power of Attorney [24-01-2025(online)].pdf 2025-01-24
14 202321067386-Form 1 (Submitted on date of filing) [24-01-2025(online)].pdf 2025-01-24
15 202321067386-Covering Letter [24-01-2025(online)].pdf 2025-01-24
16 202321067386-CERTIFIED COPIES TRANSMISSION TO IB [24-01-2025(online)].pdf 2025-01-24
17 202321067386-FORM 3 [31-01-2025(online)].pdf 2025-01-31