Abstract: Disclosed is a method and system for determining at least one combined forecast value of non-conventional energy resources. An Input/output Interface receives an adaptively selected historical dataset and a current dataset from one or more predictive forecast models and/or measurements. An adaptive forecast module generates one or more variants of machine learning models to model the performance of the one or more predictive forecast models by training the one or more variants of machine learning models on the historical dataset. The adaptive forecast module correlates the current dataset with the historical dataset to adaptively obtain a filtered historical dataset. The adaptive forecast module evaluates the one or more variants of machine learning models on the filtered historical dataset. The adaptive forecast module derives a statistical model to determine the at least one combined forecast value by combining outputs obtained based on said evaluation.
FORM 2
THE PATENTS ACT, 1970
(39 of 1970)
&
THE PATENT RULES, 2003
COMPLETE SPECIFICATION
(See Section 10 and Rule 13)
Title of invention:
METHOD AND SYSTEM FOR ADAPTIVE FORECAST OF WIND RESOURCES
Applicant:
Tata Consultancy Services Limited A company Incorporated in India under The Companies Act, 1956
Having address:
Nirmal Building, 9th Floor,
Nariman Point, Mumbai 400021.
Maharashtra. India
The following specification particularly describes the invention and the manner in which it is to be performed.
TECHNICAL FIELD
[001] The present subject matter described herein, in general, relates to data
forecasting systems, and more particularly to data forecasting systems for forecasting non-conventional energy resources using machine learning mechanisms.
BACKGROUND
[002] The depletion of conventional resources of energy has resulted in utilization of
non-conventional resources such as wind, sunlight, tides and geothermal heat etc for generating energy and power. The renewable energy generated from these natural resources plays significant role in meeting the quantum of energy required for constantly growing sectors in the global economics worldwide.
[003] There has been a tremendous growth in the utilization of wind energy for
generating power in recent times. The market analysis at the end of year 2011 depict that wind power is growing at over 20% annually, with a worldwide installed capacity of 238,000 megawatts (MW) which is majorly being used in continents such as Europe. Asia, and the United States. Considering the impact of the wind resources created in the power market while delivering quality and quantity output, accurate forecast of the wind resources is essential.
[004] There have been several attempts made in the past for accurate forecast of
wind resources. Several forecasting tools exists in the art that enable forecasting of wind resources based on different assumptions and concepts resulting in multiple alternate forecasts. Further, there have been attempts made on combining these several alternate forecasts to a single forecast of superior accuracy using various statistical and machine learning methods.
[005] One of such methods includes classifying, normalizing and grouping historic
wind patterns and associating each group with an energy output that the wind farm produces using neural network and Bayesian logic. The method uses turbine specific data, met mast data and regional forecast information from external forecasting services to adaptively adjust its logic and update current wind patterns. The wind resource is forecasted based on a match obtained from the historian database relating the updated wind pattern to wind farm energy
and use of actual measured energy and turbine specific data. However, this method is limited to using only regional forecast information and ignoring turbine level forecasts. Further, performance of a neural network changes with change in parameters governing the network; and hence use of a single neural network may not cover the entire parameter space.
[006] Another method that enables wind resource forecasting is implemented using
two sub-systems. A wind forecasting subsystem of this method adaptively combines wide area wind forecast signals, alternate meteorological data source and SCADA based inputs to produce a refined wide area wind forecast signal. This then acts as the input to another subsystem called wind farm production forecasting that uses turbine specific transfer functions and power curves to convert wide area wind forecast signals to turbine specific wind forecast signals and energy forecast respectively that is further refined based on SCADA inputs. However, the adaptive statistics module adopted by this method uses regional forecast information and ignores turbine specific forecasts.
[007] Another well-known technique disclosed in the art for estimating wind
resources forecast is a method that uses an NWP model. The NWP model in addition to receipt of wide area regional forecast as input adjusts and calibrates its forecast based on turbine level measurements. However, the NWP model may not work well for short term forecasts. Further, the NWP model receives a single forecasting service as input and hence lack combining several forecasts. In addition, this technique is limited to application to physical models only.
[008] In order to improve the accuracy of the NWP model for short term forecasting.
another method is disclosed in the art wherein multiple forecasts are generated from a single model by using slightly different initial conditions (and/or boundary conditions), which are later combined to give an ensemble forecast. However, this method uses a single NWP model; perturb the input/runtime boundary conditions to generate multiple results. The accuracy is hence bounded by the model"s performance. Further, ensemble techniques being used are mostly mathematical and hence do not involve any machine learning approach.
[009] Another approach proposed in the art involves adaptively combining alternate
forecasts by means of two methods: (1) Linearly combining them with appropriate weights assigned; (2) Exponentially weighing and tracking the best predictor. The approach further
involves selection of the best forecasts using exponential weighing. However, the weighing method has inherent limitations that it fails to adapt to changes, especially when the best predictor constantly changes.
[0010] Thus, in view of the limitations observed in the art. there is a need for a
method and system that enables adaptive forecast of wind resources by combining several alternate forecasts and achieving maximized forecast accuracy. More particularly, there is a need in the art to converge on a most efficient method of combining that is independent of the nature of predicators, whose forecasts are to be combined, applicable universally and is not limited to a specific location, and able to perform over a range of forecast time horizons.
[0011] Further, there is a need for a method and system that enables use of alternate
forecasts at turbine level. Further, there is a need for method and system that enables wind resources forecasting by creating variants of artificial neural networks that covers the entire parameter space rather than relying on single neural network.
SUMMARY
[0012] This summary is provided to introduce aspects related to systems and methods
for determining at least one combined forecast value of non-conventional energy resources and the aspects are further elaborated below in the detailed description. This Summary is not intended to identify essential features of the claimed subject matter nor is it intended for use in determining or limiting the scope of the claimed subject matter.
[0013] In one implementation, a method for determining at least one combined
forecast value of non-conventional energy resources by combining one or more forecast values received from one or more predictive forecast models is disclosed. The method , comprises a plurality of steps that are performed by a processor. The steps performed by the processor comprises: adaptively selecting a historical dataset comprising a first set of forecast values and a first set of actual values received from the one or more predictive forecast models and/or one or more measurements; dynamically generating one or more sufficiently large number of variants of machine learning models to model the performance of the one or more predictive forecast models by training the one or more variants of machine learning models on the selected historical dataset; receiving in real time a current dataset comprising a second set of forecast values and a second set of actual values from the one or more predictive forecast
models; correlating the current dataset with the historical dataset to adaptively obtain a filtered historical dataset; evaluating the one or more trained variants of machine learning models on the filtered historical dataset; and deriving a statistical model to determine at least one combined forecast value by combining the outputs of the one or more trained variants of machine learning models based on the evaluation of the one or more trained variants of machine learning models on the filtered historical dataset.
[0014] In another implementation, a system for determining at least one combined
forecast value of non-conventional energy resources by combining one or more forecast values received from one or more predictive forecast models is disclosed. The system comprises a processor, an Input/Output (I/O) interface and a memory coupled to the processor for executing a plurality of modules present in the memory. The plurality of modules comprises an adaptive forecast module. The Input/output (I/O) interface is configured to read an adaptively selected historical dataset and a current dataset received from the one or more predictive forecast models and/or measurements. The adaptive forecast module is configured to adaptively select the historical dataset comprising a first set of forecast values and a first set of actual values received from the one or more predictive forecast models and/or one or more measurements. Further, the adaptive forecast module is configured to dynamically generate one or more sufficiently large number of variants of machine learning models to model the performance of the one or more predictive forecast models by training the one or more variants of machine learning models on the historical dataset. The adaptive forecast module is then adapted to receive in real time a current dataset comprising a second set of forecast values and a second set of actual values from the one or more predictive forecast models. The adaptive forecast module is further configured to correlate the current dataset with the historical dataset to adaptively obtain a filtered historical dataset. The adaptive forecast module is further configured to evaluate the one or more trained variants of machine learning models on the filtered historical dataset. Finally, the adaptive forecast module is configured to derive a statistical model that determines the at least one combined forecast value by combining the outputs of the one or more trained variants of machine learning models based on the evaluation of the one or more trained variants of machine learning models on the filtered historical dataset.
[0015] In yet another implementation, a computer program product having embodied
thereon a computer program for determining at least one combined forecast value of non-conventional energy resources by combining one or more forecast values received from one or more predictive forecast models is disclosed. The computer program product comprises a program code for adaptively selecting a historical dataset comprising a first set of forecast values and a first set of actual values received from the one or more predictive forecast models and/or one or more measurements. The computer program code further comprises a program code for dynamically generating one or more sufficiently large number of variants of machine learning models to model the performance of the one or more predictive forecast models by training the one or more variants of machine learning models on the selected historical dataset. The computer program code further comprises a program code for receiving in real time a current dataset comprising a second set of forecast values and a second set of actual values from the one or more predictive forecast models. The computer program code further comprises a program code for correlating the current dataset with the historical dataset to adaptively obtain a filtered historical dataset. The computer program code further comprises a program code for evaluating the one or more trained variants of machine learning models on the filtered historical dataset. The computer program code further comprises a program code for deriving a statistical model to determine at least one combined forecast value by combining the outputs of the one or more trained variants of machine learning models based on the evaluation of the one or more trained variants of machine learning models on the filtered historical dataset.
BRIEF DESCRIPTION OF THE DRAWINGS
[0016] The detailed description is described with reference to the accompanying
figures. In the figures, the left-most digit(s) of a reference number identifies the figure in which the reference number first appears. The same numbers are used throughout the drawings to refer like features and components.
[0017] Figure 1 illustrates a network implementation of system for determining at
least one combined forecast value of non-conventional energy resources, in accordance with an embodiment of the present subject matter.
[0018] Figure 2 illustrates the system, in accordance with an embodiment of the
present subject matter.
[0019] Figure 3 illustrates a method for determining at least one combined forecast
value of non-convent ional energy resources, in accordance with an embodiment of the present subject matter.
[0020] Figure 4 is a flow diagram illustrating alpha-beta-gamma method of modeling
for combining forecasts, in accordance with an embodiment of the present subject matter.
[0021] Figure 5 illustrates a method, in accordance with an embodiment of the present
subject matter.
DETAILED DESCRIPTION
[0022] Systems and methods for determining at least one combined forecast value of
non-conventional energy resources are described. The present subject matter discloses an effective and efficient mechanism for determining the at least one combined forecast value by combining a plurality of forecast values received from one or more predictive forecast models. The combined forecast value is determined based on machine learning techniques in combination with Adaptive boosting (Adaboost) approach.
[0023] In order to determine the at least one combined forecast value, at first, a
historical dataset and a current dataset each comprising forecast values and actual values for the one or more predictive forecast models are received. Subsequent to the receipt of the historical dataset and the current dataset, one or more variants of machine learning models are dynamically generated to model the performance of the one or more predictive forecast models. The one or more variants of machine learning models can be generated using known ensemble techniques. In the process of dynamically generating the one or more sufficiently large number of variants of machine learning models, the one or more variants of the machine learning models are trained on the historical dataset.
[0024] Further, the historical dataset is correlated with the current dataset to
adaptively obtain a filtered historical dataset. The one or more trained variants of the machine learning models are then evaluated using the filtered historical dataset to obtain one or more outputs of the one or more trained variants of the machine learning models. The one or more
outputs are suggestive of one or more weights to be assigned to the one or more trained variants of the machine learning models and the outputs of the one or more trained variants of the machine learning models. Thus, the machine learning models are used to model the performance of the one or more predictive forecast models by comparing the forecasts values with the corresponding actual values over the historical dataset or evaluating the forecast errors of the one or more predictive forecast models over the historical dataset.
[0025] Thereafter, a statistical model is derived by adaptively assigning the one or
more weights to each of the one or more trained variants of machine learning models and the outputs of the one or more trained variants of machine learning models. The statistical model represents an optimal combination function that is utilized for determining the at least one combined forecast value. The optimal combination function determines the combined forecast value by combining the outputs of the one or more trained variants of the machine learning models using the one or more weights obtained during evaluation of the one or more variants of machine learning models. Thus, the systems and methods of the present invention enables converging on the most efficient and effective combined forecast values by adaptively combining the one or more forecast values obtained through the one or more predictive forecast models.
[0026] While aspects of described system and method for determining at least one
combined forecast value of non-conventional energy resources may be implemented in any number of different computing systems, environments, and/or configurations, the embodiments are described in the context of the following exemplary system.
[0027] Referring now to Figure 1, a network implementation 100 of a system 102 for
determining at least one combined forecast value of non-conventional energy resources is illustrated, in accordance with an embodiment of the present subject matter. In one embodiment, the system 102 enables receiving of a historical dataset comprising historical forecast values along with corresponding actual values therewith and a current dataset comprising current forecast values along with corresponding actual values therewith from one or more predictive forecast models. The system 102 dynamically generates one or more variants of machine learning models for modeling the performance of the one or more predictive forecast models by training the one or more variants of the machine learning
models on the historical dataset. The system 102 may correlate the historical dataset with the current dataset to obtain a filtered historical dataset. The system 102 may then evaluate the trained one or more variants of the machine learning models using the filtered historical dataset. Based upon the evaluation, one or more outputs in form of one or more weights may be obtained. The system 102 may then derive a statistical model representing an optimal combination function by adaptively assigning the one or more weights to the trained one or more variants of the machine learning models and/or the outputs of each of the trained one or more variants of the machine learning models. The system 102 may determine the combined forecast value using the optimal combination function, wherein the optimal combination function adaptively combines the one or more weights assigned to the one or more trained variants of the machine learning models and the outputs of each of the trained one or more variants of the machine learning models.
[0028] Although the present subject matter is explained considering that the system
102 is implemented on a server, it may be understood that the system 102 may also be implemented in a variety of computing systems, such as a laptop computer, a desktop computer, a notebook, a workstation, a mainframe computer, a server, a network server, and the like. It will be understood that the system 102 may be accessed by multiple users through one or more user devices 104-1, 104-2... 104-N, collectively referred to as user devices 104 hereinafter, or applications residing on the user devices 104. Examples of the user devices 104 may include, but are not limited to, a portable computer, a personal digital assistant, a handheld device, and a workstation. The user devices 104 are communicatively coupled to the system 102 through a network 106.
[0029] In one implementation, the network 106 may be a wireless network, a wired
network or a combination thereof. The network 106 can be implemented as one of the different types of networks, such as intranet, local area network (LAN), wide area network (WAN), the internet, and the like. The network 106 may either be a dedicated network or a shared network. The shared network represents an association of the different types of networks that use a variety of protocols, for example, Hypertext Transfer Protocol (HTTP), Transmission Control Protocol/Internet Protocol (TCP/IP), Wireless Application Protocol (WAP), and the like, to communicate with one another. Further the network 106 may include
a variety of network devices, including routers, bridges, servers, computing devices, storage devices, and the like.
[0030] Referring now to Figure 2, the system 102 is illustrated in accordance with an
embodiment of the present subject matter. In one embodiment, the system 102 may include at least one processor 202, an input/output (I/O) interface 204. and a memory 206. The at least one processor 202 may be implemented as one or more microprocessors, microcomputers, microcontrollers, digital signal processors, central processing units, state machines, logic circuitries, and/or any devices that manipulate signals based on operational instructions. Among other capabilities, the at least one processor 202 is configured to fetch and execute computer-readable instructions stored in the memory 206.
[0031] The I/O interface 204 may include a variety of software and hardware
interfaces, for example, a web interface, a graphical user interface, and the like. The I/O interface 204 may allow the system 102 to interact with a user directly or through the user devices 104. Further, the I/O interface 204 may enable the system 102 to communicate with other computing devices, such as web servers and external data servers (not shown). The I/O interface 204 can facilitate multiple communications within a wide variety of networks and protocol types, including wired networks, for example, LAN, cable, etc., and wireless networks, such as WLAN, cellular, or satellite. The I/O interface 204 may include one or more ports for connecting a number of devices to one another or to another server.
[0032] The memory 206 may include any computer-readable medium or computer
program product known in the art including, for example, volatile memory, such as static random access memory (SRAM) and dynamic random access memory (DRAM), and/or non¬volatile memory, such as read only memory (ROM), erasable programmable ROM, flash memories, hard disks, optical disks, and magnetic tapes. The memory 206 may include modules 208 and data 210.
[0033] The modules 208 include routines, programs, objects, components, data
structures, etc., which perform particular tasks or implement particular abstract data types. In one implementation, the modules 208 may include an adaptive forecast module 212 and other modules 214. The other modules 214 may include programs or coded instructions that supplement applications and functions of the system 102.
[0034] The data 210, amongst other things, serves as a repository for storing data
processed, received, and generated by one or more of the modules 208. The data 210 may also include a predictive forecast database 216, a refined forecast database 218. a system database 220, and other data 222. The other data 222 may include data generated as a result of the execution of one or more modules in the other module 214.
[0035] In one implementation, at first, a user may use the user devices 104 to access
the system 102 via the I/O interface 204. The user may register using the I/O interface 204 in order to use the system 102. The working of the system 102 may be explained in detail in Figures 3 and 4 explained below.
[0036] Referring to Figure 3, a detailed working of the adaptive forecast module 212
along with the working of other components of the system 102 is illustrated, in accordance with an embodiment of the present subject matter. In one implementation, in-order to determine the at least one combined forecast value, a plurality of forecast values and corresponding actual values are read by the I/O Interface 204 from a plurality of predictive forecast models 302-1, 302-2....302-N. In one embodiment, the plurality of predictive forecast models 302-1, 302-2....302-N comprise a supervisory control and data acquisition (SCADA) model, physical model such as numerical weather prediction model, statistical model, machine learning model, and an alternate forecast model etc. The plurality of forecast values and the corresponding actual values are stored in the predictive forecast database 216. The predictive forecast database 216 is adapted to store a historical dataset comprising the plurality of forecast values and the corresponding actual values received from the plurality of predictive forecast models 302-1, 302-2....302-N historically. Further, the predictive forecast database 216 is adapted to store a current dataset historical dataset comprising the plurality of forecast values and the corresponding actual values received from the plurality of predictive forecast models 302-1, 302-2....302-N in real time.
[0037] In one implementation, the adaptive forecast module 212 may receive the
historical dataset and the current dataset stored in the predictive forecast database 216 using a receiving module 304. A machine learning module 306 is configured for generation of one or more variants of machine learning models to model the performance of the plurality of predictive forecast models 302-1, 302-2....302-N. In one embodiment, the one or more
variants of machine learning models generated comprise Artificial Neural Networks (ANNs), basis function models, kernel methods, support vector machines, decision trees, variation methods, distribution sampling methods, ensemble methods, graphical models, and search methods etc. for mining records. The one or more variants of machine learning models are generated using ensemble techniques comprising bagging, boosting, adaptive boosting (AdaBoost), stack generalization. Bayesian model combination, clustering methods, tree based models and conditional mixture models etc. In a preferred embodiment, the one or more variants are generated using the AdaBoost technique.
[0038] Further, the machine learning module 306 is configured for training the one or
more variants of machine learning models on the historical dataset received. The one or more variants of the machine learning models are trained using AdaBoost technique. A set of training data points from the historical dataset is introduced and assigned equal weights. The starting values of the model parameters are initialized. On the entire set of training data points, a machine learning model in form of variant 1 (also referred as Model 1) is fitted. All the training data points within a certain Euclidean distance from this model are modeled by the Model 1. Further, the data points modeled by the Model 1 are removed, and the remaining data points are re-weighed with respect to their Euclidean distance from the Model 1 with the points farther from the Model 1 gaining more weights and vice versa. The new re-weighted training set containing the data points that were un- modeled by the Model 1 is used for training another model called variant 2 or a Model 2. All the data points modeled by the Model 2 are removed from the re-weighted data set and the rest of the data points are weighed with respect to their Euclidean distance from the Model 2. This process continues till all the data points in the set of training data points are modeled by one or the other model variants. Thus, the one or more variants of machine learning models are generated and trained for the plurality of predictive forecast models 302-1, 302-2.... 302-N.
[0039] In one embodiment, a correlation module 310 is utilized to correlate the
historical dataset with the current dataset to adaptively obtain a filtered historical dataset. The correlation module 310 is configured for correlating the current forecast value with the historical forecast value. Such correlation results in obtaining the filtered historical dataset. wherein the filtered historical dataset comprises of the historical forecast values highly correlated to the current forecast value along with their corresponding historical actual values.
The filtered historical dataset is then inputted to an evaluation module 312. Further, the evaluation module 312 is configured to evaluate the performance of the one or more variants of machine learning models trained over the adaptively obtained filtered dataset. The evaluation module as a result of evaluation is configured to adaptively assign weights to each of the trained variants of machine learning models and to their output values.
[0040] Subsequent to adaptive assignment of weights to each of the trained variants of
machine learning models, a statistical analysis module 314 is configured to derive a statistical model for adaptively combining the weights assigned to each of the trained variants of machine learning models and their outputs. In an embodiment, the statistical model represents an optimal combination function that is responsible for adaptively combining the weights assigned to each of the trained variants of machine learning models and their outputs. The optimal combination function based on the combination of weights results in obtaining the at least one combined forecast value for the one or more predictive forecast models 302-1, 302-2....302-N. Thus, the output of the statistical analysis module 314 is the at least one combined forecast value for the one or more predictive forecast models 302-1, 302-2....302-N. The I/O interface 204 is configured to write the at least one combined forecast value in the refined forecast database 218 for future use. Since the one or more trained variants of machine learning models and their output values are adaptively combined, the at least one forecast value represents an accurate forecast value for the one or more predictive forecast models 302-1, 302-2....302-N.
[0041] In one exemplary embodiment, the at least one combined forecast value can be
obtained using AdaBoost technique by implementing an alpha-beta-gamma (αβγ) algorithm. In this embodiment, for example, consider k number of readily available forecasts fl,f2.f3...fk, of the random variable z, represented by a discrete time series {...zt-5,zt-4,zt-3,zt-2,zt-l .zt}, the objective is to combine them into a single forecast fc such that fc is a better forecast than any of the readily available forecasts. The combination can be represented in the following format:
fc(z) = o(fl(z).f2(z)3(z) ....fk(z)) (1)
Where, Φ is a non-linear function of the k forecasts f1 to fk. The non-linear relationship between the individual predictors can be justified in the following manner:
[0042] In this embodiment, assume that the information set used by the jth predictor
for the ith individual forecast is given as {Iji: Ijci. Ijsi}, where Ijci is the common part of information used by all the k predictors and 1jsi is the special information for the ith forecast only. The combination model can be considered as a system that combines information processing sub-systems given as:
fc = Fc (I1.12.13...Ik} (2)
[0043] The performance of the integrated system is more than just a linear sum of
performances of individual subsystems and hence the non-linear relationship between individual forecasts and the actual time series represented by equation (1) is justified. It is generally difficult to determine the form of the non-linear relationship "O". However, using data driven forecasting procedures using machine learning models; the non-linear relationship can be realized.
[0044] In this embodiment, the historical dataset and the current dataset is accepted
from the Predictive forecast Database 216 by implementing four different statistical forecast algorithms. The time series data is first analyzed and forecasted using Mycielski approach, Persistence. Modified Persistence and Artificial Neural Network. The forecast values and the actual values of the time series are arranged as columns of a Data Matrix (DM), with columns 1 to 4 representing the forecast values and column 5 containing the actual values of the time series. In other embodiments, there may be other methods or ways of arranging the forecast values and the actual values other than the Data matrix (DM) method. In this embodiment, the Data matrix formed is as below:
[0045] Subsequent to the formation of the Data matrix, the alpha-beta-gamma (αβγ)
algorithm is used to map the values of the four forecasts in column 1 to column 4 of the data matrix to the actual values of wind speed in column 5, nonlinearly. The complexity of the artificial neural networks (ANN) variants is determined using a number of techniques such as trial and errors, signal to noise ratio, ROC curves etc. The variants of artificial neural networks are generated using various ensemble techniques such as bagging, boosting, adaptive boosting (AdaBoost), stack generalization, Bayesian model combination, clustering methods, tree based models and conditional mixture models etc. Preferably, adaptive boosting (AdaBoost) is utilized for generating the variants of artificial neural networks (ANN). Generally, the performance of Artificial Neural Networks (ANNs) changes with the change in internal parameters governing the network. Even with the same set of parameters, the ANN results might vary significantly. Hence creation of variants of ANNs is vital for each parameter set over a number of runs as well as across the entire Parameter Space (PS). Therefore, a smartly modified version of the AdaBoost algorithm is utilized to accomplish the creation of ANN variants. This novel modification is necessary due to the notoriously varying nature of the wind resource as well as the rapidly fluctuating performances of the individual forecasts. Since the data represents a time series, the ordered structure of Training Set (TRS), Validation Set (VLS) and Forecasting Set (FDS) are retained in the alpha-beta-gamma (αβγ) algorithm. In this exemplary embodiment the final model combines the forecasts by filtering the best possible combinations throughout the parameter space and recombining them to obtain a single better forecast for each time instance. Figure 4 illustrates a flow diagram collectively describing the alpha-beta-gamma (αβγ) algorithmic method of modeling to obtain a single forecast by combining one or more forecasts in one embodiment of the invention.
[0046] As illustrated, at step 402, the data matrix is formed by initialization of the
training data set (TRS), the validation data set (VLS), and the forecasting data set (FDS)
[0047] At step 404, the variants of artificial neural networks (ANN) are generated
using AdaBoost algorithm by training on the training data set (TRS). In this embodiment, "i" variants of ANN are generated using the AdaBoost algorithm by training on the training data set (TRS). The training on the TRS is repeated for "j:" times on the current parameter set (by varying the parameter values internal to the ANNs so that a number of variants with different
ending points are generated). The variants generated are stored as data structure "a" in the system database 220.
[0048] At step 406. it is verified whether the parameter space is covered sufficiently.
If the parameter space is covered sufficiently, the method is proceeded to step 408, else the parameters are updated and the step 404 is re-executed.
[0049] At step 408, the forecast data set (FDS) and the validation set (VLS) are
correlated.
|0050] At step 410, the data points from the validation data set (VLS) are filtered
(VLF) with respect to the current FDS data points based on the correlations from step 408. Then the performance of ANN variants is validated on the current VLF data points. The performance values are stored in matrix "β" in the system database 220. In alternative embodiment, the performance values may be stored in data formats/types other than the matrix "β".
[0051] At step 412, the combination of ANN variants is implemented to obtain a
resulting combination model "γ" in a manner such that y= f (αs, βs), wherein, βs represents a set of best performance values and as represents a set of best performing variants of ANN. The resulting combination model "γ" is then stored in the system database 220.
[0052] At step 414, the resulting combination model (y) is utilized for making
predictions on the current FDS data points.
[0053] The execution of "α β γ" algorithm on the Data Matrix is experimentally tested
to validate the performance of the "α β γ" algorithm. The performance of "α β γ" is compared with four individual forecasts in terms of Mean Absolute Error (MAE), Mean Absolute Percentage Error (MAPE) and R-square (R2). The following table illustrates the result of comparison of performances of the four predictors and "α β γ" over a forecast period of one year. The algorithm "α β γ" is tested to predict wind speeds for 1 data point ahead, 24 data points ahead and 48 data points ahead forecast horizons.
M S P ANN α β γ Forecast horizon
1 α β γ Forecast horizon
24 α β γ Forecast horizon
48
MAPE 17.98123 17.34337 20.82771 7.964045 5.621898 5.671842 5.723878
R2 0.894285 0.889254 0.836484 0.989425 0.988417 0.987741 0.985565
MAE 0.6931 0.71497 0.8997 0.22 0.1798 0.1588 0.16894
[0054] It is observed from the table that the performance of "α β γ" is independent of
the forecast horizon. In addition, less complex ANNs performed better compared to ANNs with more number of hidden layers and more number of neurons in each layer. Hence, the processing time required for forecast over various windows is less or almost the same.
[0055] Referring now to Figure 5, a method 500 for determining at least one
combined forecast value of non-conventional energy resources is shown, in accordance with an embodiment of the present subject matter. The method 500 may be described in the general context of computer executable instructions. Generally, computer executable instructions can include routines, programs, objects, components, data structures, procedures, modules, functions, etc., that perform particular functions or implement particular abstract data types. The method 500 may also be practiced in a distributed computing environment where functions are performed by remote processing devices that are linked through a communications network. In a distributed computing environment, computer executable instructions may be located in both local and remote computer storage media, including memory storage devices.
[0056] The order in which the method 500 is described is not intended to be construed
as a limitation, and any number of the described method blocks can be combined in any order to implement the method 500 or alternate methods. Additionally, individual blocks may be deleted from the method 500 without departing from the spirit and scope of the subject matter described herein. Furthermore, the method can be implemented in any suitable hardware, software, firmware, or combination thereof. However, for ease of explanation, in the
embodiments described below, the method 500 may be considered to be implemented in the above described system 102.
[0057] At block 502, a historical dataset comprising a first set of forecast values and a
first set of actual values is received. In one implementation, the historical dataset may be received by the receiving module 304 of the Adaptive Forecast Module 212.
[0058] At block 504. one or more variants of machine learning models to model the
performance of the one or more predictive forecast models are generated by training the one or more variants of machine learning models on the historical dataset. In one implementation, the one or more variants of machine learning models may be generated and trained by the machine learning module 306 of the Adaptive Forecast Module 212.
[0059] At block 508. a current dataset comprising a second set of forecast values and a
second set of actual values is received. In one implementation, the current dataset may be received by the receiving module 304 of the Adaptive Forecast Module 212.
[0060] At block 510, the current dataset is correlated with the historical dataset in
order to adaptively obtain a filtered historical dataset. In one implementation, the current dataset and the historical dataset may be adaptively correlated by the correlation module 310 of the Adaptive Forecast Module 212 to obtain the filtered historical dataset.
[0061] At block 512, the trained one or more variants of machine learning models are
evaluated over the filtered historical dataset in order to assigning weights to each of the trained one or more variants of machine learning models and their outputs. In one implementation, the trained one or more variants of machine learning models may be evaluated by the evaluation module 312 of the Adaptive Forecast Module 212.
[0062] At block 514. the output of the evaluation module 312, i.e. the weights
assigned may be combined to derive a statistical model in form of an adaptive combination function. The adaptive combination function combines the weights assigned to each of the trained one or more variants of machine learning models and their outputs in order to obtain the at, least one combined forecast value. In one implementation, the statistical model may be derived by the statistical analysis module 314 of the Adaptive Forecast Module 212.
ADVANTAGES OF THE INVENTION
[0063] The present invention enables combining of one or more alternate forecasts
that are independent of the nature of the predictors whose forecasts are to be combined,
[0064] The present invention enables generation of variants of neural networks that
cover the entire parameter space and which are trained on historical and real-time data set of alternate forecasts.
[0065] The present invention adopts an adaptive method for wind resources
forecasting using advanced machine learning techniques that enables maximized forecast accuracy.
[0066] The optimal combination function can be re-established for every single
forecast based on the observed alternate forecasts.
[0067] Although implementations for methods and systems for determining at least
one combined forecast value of non-conventional energy resources have been described in language specific to structural features and/or methods, it is to be understood that the appended claims are not necessarily limited to the specific features or methods described. Rather, the specific features and methods are disclosed as examples of implementations for determining at least one combined forecast value of non-conventional energy resources.
[0068] The present invention takes informed risk of predicting closer towards the
sample rather than the mean of the predictive distribution. Thus, the errors are minimized even more compared to other alternative forecast models to be combined and hence out performs.
WE CLAIM:
1. A method for determining at least one combined forecast value of non-conventional energy resources by combining one or more forecast values received from one or more predictive forecast models, the method comprising:
adaptively selecting, by the processor, a historical dataset comprising a first set of forecast values and a first set of actual values received from the one or more predictive forecast models and/or one or more measurements;
generating, by a processor, one or more variants of machine learning models to model a performance of the one or more predictive forecast models by training the one or more variants of machine learning models on the historical dataset;
receiving in real time, by the processor, a current dataset comprising a second set of forecast values and a second set of actual values from the one or more predictive forecast models;
correlating, by the processor, the current dataset with the historical dataset to adaptively obtain a filtered historical dataset;
evaluating, by the processor, the one or more variants of machine learning models trained on the filtered historical dataset; and
deriving, by the processor, a statistical model to determine at least one combined forecast value by combining outputs of the one or more variants of machine learning models trained based on the evaluating of the one or more variants of machine learning models on the filtered historical dataset.
2. The method of claim 1, wherein the machine learning models are used to model the performance of the one or more predictive forecast models by comparing the respective forecasts values with the corresponding actual values over the historical dataset, or evaluating the forecast errors of the one or more predictive forecast models over the historical dataset.
3. The method of claim 1, wherein the one or more predictive forecast models comprises of a supervisory control and data acquisition (SCADA) model, physical model including
numerical weather prediction model, statistical model, machine learning model, an alternate forecast model or combinations thereof.
4. " The method of claim 1, wherein the one or more variants of machine learning models
comprises of Artificial Neural "Networks (ANNs), basis function models, kernel methods,
support vector machines, decision trees, variation methods, distribution sampling methods,
ensemble methods, graphical models, search methods or combinations thereof.
5. The method of claim 1, wherein the one or more variants are generated using ensemble techniques comprising bagging, boosting, AdaBoost, stack generalization, Bayesian model combination, clustering methods, tree based models, conditional mixture models or combinations thereof.
6. The method of claim 1, wherein the statistical model represents an optimal combination function which is derived by adaptively assigning a weight to each of the one or more variants of machine learning models trained and its outputs.
7. The method of claim 6, wherein the optimal combination function determines the at least one combined forecast value by combining the outputs of the one or more variants of machine learning models trained using weights obtained during evaluation of the one or more variants of machine learning models trained.
8. A system for determining at least one combined forecast value of non-conventional energy resources by combining one or more forecast values received from one or more predictive forecast models, the system comprising:
a processor (202);
an I/O interface (204) configured to read an adaptively selected historical dataset and a current dataset received from the one or more predictive forecast models and/or measurements, wherein the I/O interface (204) is further configured to write the at least one combined forecast value; and
a memory (206) coupled to the processor, wherein the processor (202) is capable of executing a plurality of modules (208) and data (210) stored in the memory, and wherein the plurality of modules (208) comprising:
an adaptive forecast module (212) configured for:
adaptively selecting a historical dataset comprising a first set of forecast values and a first set of actual values received from the one or more predictive forecast models and/or one or more measurements;
generating one or more variants of machine learning models to model a performance of the one or more predictive forecast models by training the one or more variants of machine learning models on the historical dataset;
receiving in real time a current dataset comprising a second set of forecast values and a second set of actual values from the one or more predictive forecast models;
correlating the current dataset with the historical dataset to adaptively obtain a filtered historical dataset;
evaluating the one or more variants of machine learning models trained on the filtered historical dataset;
deriving a statistical model to determine at least one combined forecast value by combining outputs of the one or more variants of machine learning models based on the evaluating of the one or more variants of machine learning models trained on the filtered historical dataset; the data (210) comprising:
a predictive forecast database configured for storing the historical dataset and the current dataset;
a refined forecast database configured for storing the at least one combined forecast value.
9. The system of claim 8, wherein the one or more predictive forecast models comprises of a supervisory control and data acquisition (SCADA) model, physical model including
numerical weather prediction model, statistical model, machine learning model, an alternate forecast model or combinations thereof.
10. The system of claim 8, wherein the one or more variants of machine learning models comprises of Artificial Neural Networks (ANNs), basis function models, kernel methods, support vector machines, decision trees, variation methods, distribution sampling methods, ensemble methods, graphical models or combinations thereof.
11. The system of claim 8, wherein the one or more variants are generated using ensemble techniques comprising bagging, boosting, AdaBoost, stack generalization, Bayesian model combination, clustering methods, tree based models, conditional mixture models or combinations thereof.
12. The system of claim 8, wherein the statistical model represents an optimal combination function which is derived by adaptively assigning a weight to each of the one or more variants of machine learning models trained and its outputs.
13. The system of claim 12, wherein the optimal combination function determines the at least one combined forecast value by combining the weights assigned to each of the one or more variants of machine learning models trained.
14. A computer program product having embodied thereon a computer program for determining at least one combined forecast value of non-conventional energy resources by combining one or more forecast values received from one or more predictive forecast models, the computer program product comprising:
a program code for adaptively selecting a historical dataset comprising a first set of forecast values and a first set of actual values received from the one or more predictive forecast models and/or one or more measurements;
a program code for generating one or more variants of machine learning models to model a performance of the one or more predictive forecast models by training the one or more variants of machine learning models on the historical dataset;
a program code for receiving in real time a current dataset comprising a second set of forecast values and a second set of actual values from the one or more predictive forecast models;
a program code for correlating the current dataset with the historical dataset to adaptively obtain a filtered historical dataset;
a program code for evaluating the one or more variants of machine learning models trained on the filtered historical dataset; and
a program code for deriving a statistical model to determine at least one combined forecast value by combining outputs of the one or more variants of machine learning models trained based on the evaluating of the one or more variants of machine learning models trained on the filtered historical dataset.
| Section | Controller | Decision Date |
|---|---|---|
| # | Name | Date |
|---|---|---|
| 1 | 2092-MUM-2012-IntimationOfGrant19-01-2023.pdf | 2023-01-19 |
| 1 | Form 3 [21-12-2016(online)].pdf | 2016-12-21 |
| 2 | 2092-MUM-2012-PatentCertificate19-01-2023.pdf | 2023-01-19 |
| 2 | ABSTRACT1.jpg | 2018-08-11 |
| 3 | 2092-MUM-2012-Written submissions and relevant documents [16-01-2023(online)].pdf | 2023-01-16 |
| 3 | 2092-MUM-2012-FORM 5(28-3-2013).pdf | 2018-08-11 |
| 4 | 2092-MUM-2012-PETITION UNDER RULE 137 [13-01-2023(online)].pdf | 2023-01-13 |
| 4 | 2092-MUM-2012-FORM 3(28-3-2013).pdf | 2018-08-11 |
| 5 | 2092-MUM-2012-RELEVANT DOCUMENTS [13-01-2023(online)].pdf | 2023-01-13 |
| 5 | 2092-MUM-2012-FORM 2[TITLE PAGE].pdf | 2018-08-11 |
| 6 | 2092-MUM-2012-FORM-26 [29-12-2022(online)]-1.pdf | 2022-12-29 |
| 6 | 2092-MUM-2012-FORM 26(30-7-2012).pdf | 2018-08-11 |
| 7 | 2092-MUM-2012-FORM-26 [29-12-2022(online)].pdf | 2022-12-29 |
| 7 | 2092-MUM-2012-FORM 2.pdf | 2018-08-11 |
| 8 | 2092-MUM-2012-FORM 2(TITLE PAGE)-(28-3-2013).pdf | 2018-08-11 |
| 8 | 2092-MUM-2012-Correspondence to notify the Controller [28-12-2022(online)].pdf | 2022-12-28 |
| 9 | 2092-MUM-2012-FORM 2(28-3-2013).pdf | 2018-08-11 |
| 9 | 2092-MUM-2012-US(14)-HearingNotice-(HearingDate-04-01-2023).pdf | 2022-12-19 |
| 10 | 2092-MUM-2012-CLAIMS [17-01-2020(online)].pdf | 2020-01-17 |
| 10 | 2092-MUM-2012-FORM 18(28-3-2013).pdf | 2018-08-11 |
| 11 | 2092-MUM-2012-COMPLETE SPECIFICATION [17-01-2020(online)].pdf | 2020-01-17 |
| 11 | 2092-MUM-2012-FORM 13(14-8-2012).pdf | 2018-08-11 |
| 12 | 2092-MUM-2012-FER_SER_REPLY [17-01-2020(online)].pdf | 2020-01-17 |
| 12 | 2092-MUM-2012-FORM 1.pdf | 2018-08-11 |
| 13 | 2092-MUM-2012-FORM 1(7-12-2012).pdf | 2018-08-11 |
| 13 | 2092-MUM-2012-OTHERS [17-01-2020(online)].pdf | 2020-01-17 |
| 14 | 2092-MUM-2012-FER.pdf | 2019-07-17 |
| 14 | 2092-MUM-2012-FORM 1(14-8-2012).pdf | 2018-08-11 |
| 15 | 2092-MUM-2012-ABSTRACT(28-3-2013).pdf | 2018-08-11 |
| 15 | 2092-MUM-2012-DRAWING.pdf | 2018-08-11 |
| 16 | 2092-MUM-2012-ABSTRACT.pdf | 2018-08-11 |
| 16 | 2092-MUM-2012-DRAWING(28-3-2013).pdf | 2018-08-11 |
| 17 | 2092-MUM-2012-DESCRIPTION(PROVISIONAL).pdf | 2018-08-11 |
| 17 | 2092-MUM-2012-CLAIMS(28-3-2013).pdf | 2018-08-11 |
| 18 | 2092-MUM-2012-CORRESPONDENCE(14-8-2012).pdf | 2018-08-11 |
| 18 | 2092-MUM-2012-DESCRIPTION(COMPLETE)-(28-3-2013).pdf | 2018-08-11 |
| 19 | 2092-MUM-2012-CORRESPONDENCE(28-3-2013).pdf | 2018-08-11 |
| 19 | 2092-MUM-2012-CORRESPONDENCE.pdf | 2018-08-11 |
| 20 | 2092-MUM-2012-CORRESPONDENCE(30-7-2012).pdf | 2018-08-11 |
| 20 | 2092-MUM-2012-CORRESPONDENCE(7-12-2012).pdf | 2018-08-11 |
| 21 | 2092-MUM-2012-CORRESPONDENCE(30-7-2012).pdf | 2018-08-11 |
| 21 | 2092-MUM-2012-CORRESPONDENCE(7-12-2012).pdf | 2018-08-11 |
| 22 | 2092-MUM-2012-CORRESPONDENCE(28-3-2013).pdf | 2018-08-11 |
| 22 | 2092-MUM-2012-CORRESPONDENCE.pdf | 2018-08-11 |
| 23 | 2092-MUM-2012-CORRESPONDENCE(14-8-2012).pdf | 2018-08-11 |
| 23 | 2092-MUM-2012-DESCRIPTION(COMPLETE)-(28-3-2013).pdf | 2018-08-11 |
| 24 | 2092-MUM-2012-DESCRIPTION(PROVISIONAL).pdf | 2018-08-11 |
| 24 | 2092-MUM-2012-CLAIMS(28-3-2013).pdf | 2018-08-11 |
| 25 | 2092-MUM-2012-ABSTRACT.pdf | 2018-08-11 |
| 25 | 2092-MUM-2012-DRAWING(28-3-2013).pdf | 2018-08-11 |
| 26 | 2092-MUM-2012-ABSTRACT(28-3-2013).pdf | 2018-08-11 |
| 26 | 2092-MUM-2012-DRAWING.pdf | 2018-08-11 |
| 27 | 2092-MUM-2012-FER.pdf | 2019-07-17 |
| 27 | 2092-MUM-2012-FORM 1(14-8-2012).pdf | 2018-08-11 |
| 28 | 2092-MUM-2012-FORM 1(7-12-2012).pdf | 2018-08-11 |
| 28 | 2092-MUM-2012-OTHERS [17-01-2020(online)].pdf | 2020-01-17 |
| 29 | 2092-MUM-2012-FER_SER_REPLY [17-01-2020(online)].pdf | 2020-01-17 |
| 29 | 2092-MUM-2012-FORM 1.pdf | 2018-08-11 |
| 30 | 2092-MUM-2012-COMPLETE SPECIFICATION [17-01-2020(online)].pdf | 2020-01-17 |
| 30 | 2092-MUM-2012-FORM 13(14-8-2012).pdf | 2018-08-11 |
| 31 | 2092-MUM-2012-CLAIMS [17-01-2020(online)].pdf | 2020-01-17 |
| 31 | 2092-MUM-2012-FORM 18(28-3-2013).pdf | 2018-08-11 |
| 32 | 2092-MUM-2012-FORM 2(28-3-2013).pdf | 2018-08-11 |
| 32 | 2092-MUM-2012-US(14)-HearingNotice-(HearingDate-04-01-2023).pdf | 2022-12-19 |
| 33 | 2092-MUM-2012-Correspondence to notify the Controller [28-12-2022(online)].pdf | 2022-12-28 |
| 33 | 2092-MUM-2012-FORM 2(TITLE PAGE)-(28-3-2013).pdf | 2018-08-11 |
| 34 | 2092-MUM-2012-FORM 2.pdf | 2018-08-11 |
| 34 | 2092-MUM-2012-FORM-26 [29-12-2022(online)].pdf | 2022-12-29 |
| 35 | 2092-MUM-2012-FORM 26(30-7-2012).pdf | 2018-08-11 |
| 35 | 2092-MUM-2012-FORM-26 [29-12-2022(online)]-1.pdf | 2022-12-29 |
| 36 | 2092-MUM-2012-RELEVANT DOCUMENTS [13-01-2023(online)].pdf | 2023-01-13 |
| 36 | 2092-MUM-2012-FORM 2[TITLE PAGE].pdf | 2018-08-11 |
| 37 | 2092-MUM-2012-PETITION UNDER RULE 137 [13-01-2023(online)].pdf | 2023-01-13 |
| 37 | 2092-MUM-2012-FORM 3(28-3-2013).pdf | 2018-08-11 |
| 38 | 2092-MUM-2012-Written submissions and relevant documents [16-01-2023(online)].pdf | 2023-01-16 |
| 38 | 2092-MUM-2012-FORM 5(28-3-2013).pdf | 2018-08-11 |
| 39 | ABSTRACT1.jpg | 2018-08-11 |
| 39 | 2092-MUM-2012-PatentCertificate19-01-2023.pdf | 2023-01-19 |
| 40 | Form 3 [21-12-2016(online)].pdf | 2016-12-21 |
| 40 | 2092-MUM-2012-IntimationOfGrant19-01-2023.pdf | 2023-01-19 |
| 41 | 2092-MUM-2012-FORM 4 [27-09-2025(online)].pdf | 2025-09-27 |
| 1 | serachTpo_25-06-2019.pdf |