Next Article in Journal
Knowledge-Enhanced Deep Learning for Simulation of Extratropical Cyclone Wind Risk
Previous Article in Journal
Retrieval of High-Resolution Aerosol Optical Depth for Urban Air Pollution Monitoring
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Framework Based on the Stacking Ensemble Machine Learning (SEML) Method: Application in Wind Speed Modeling

by
Amirreza Morshed-Bozorgdel
,
Mojtaba Kadkhodazadeh
,
Mahdi Valikhan Anaraki
and
Saeed Farzin
*
Department of Water Engineering and Hydraulic Structures, Faculty of Civil Engineering, Semnan University, Semnan 35131-19111, Iran
*
Author to whom correspondence should be addressed.
Submission received: 12 April 2022 / Revised: 27 April 2022 / Accepted: 6 May 2022 / Published: 8 May 2022
(This article belongs to the Section Atmospheric Techniques, Instruments, and Modeling)

Abstract

:
Wind speed (WS) is an important factor in wind power generation. Because of this, drastic changes in the WS make it challenging to analyze accurately. Therefore, this study proposed a novel framework based on the stacking ensemble machine learning (SEML) method. The application of a novel framework for WS modeling was developed at sixteen stations in Iran. The SEML method consists of two levels. In particular, eleven machine learning (ML) algorithms in six categories neuron based (artificial neural network (ANN), general regression neural network (GRNN), and radial basis function neural network (RBFNN)), kernel based (least squares support vector machine-grid search (LSSVM-GS)), tree based (M5 model tree (M5), gradient boosted regression (GBR), and least squares boost (LSBoost)), curve based (multivariate adaptive regression splines (MARS)), regression based (multiple linear regression (MLR) and multiple nonlinear regression (MNLR)), and hybrid algorithm based (LSSVM-Harris hawks optimization (LSSVM-HHO)) were selected as the base algorithms in level 1 of the SEML method. In addition, LSBoost was used as a meta-algorithm in level 2 of the SEML method. For this purpose, the output of the base algorithms was used as the input for the LSBoost. A comparison of the results showed that using the SEML method in WS modeling greatly affected the performance of the base algorithms. The highest correlation coefficient (R) in the WS modeling at the sixteen stations using the SEML method was 0.89. The SEML method increased the WS modeling accuracy by >43%.

1. Introduction

Wind energy is one of the most important renewable energy sources due to the fact of its stability and nonpolluting and free nature [1]. This is why wind energy can help generate electricity. Accurate modeling and WS prediction are critical to management and planning [2]. However, WS modeling and prediction is complex due to the large fluctuations of this variable. Therefore, providing novel methods for WS analysis can help solve this problem. One of the most important and newest tools in modeling and predicting WS is ML algorithms.
In recent years, artificial intelligence (AI) algorithms to solve various problems have expanded rapidly [3,4,5]. Since modeling and predicting different parameters are challenging and costly today, algorithms have made the job easier [6,7]. The advantages of AI algorithms include easy implementation, low cost, and high accuracy. Increasing the accuracy of algorithms in modeling and predicting various parameters is an important issue. For this reason, attempts are being made to improve the accuracy of ML algorithms using novel methods.
In recent years, classical algorithms such as ANN, adaptive neuro-fuzzy inference system (ANFIS), support vector machine (SVM), LSSVM, twin support vector regression (TSVR), random forest regression (RFR), and convolutional neural networks (CNNs) have had many applications in modeling and predicting WS [8,9,10]. Elsisi et al. [11] used ANFIS to analyze WS data. In addition, hybrid algorithms obtained from simulations and optimization algorithms have been used to model and predict WS data [12,13,14]. Du et al. [15] proposed a new hybrid algorithm based on multi-objective optimization. The results showed that the performance of the new hybrid algorithm was acceptable. Similarly, in another study, Cheng et al. [16] proposed a new hybrid algorithm based on multi-objective salp swarm optimization (MSSO). The results showed that the new hybrid algorithm had good performance in predicting WS. Another study used a new hybrid algorithm, called ensemble empirical mode decomposition genetic algorithm long short-term memory (EEMD-GA-LSTM), to predict WS [17]. The results showed that the new hybrid algorithm had high potential for predicting WS. In another group of studies, researchers used deep learning to model and predict WS [18,19]. Neshat et al. [20] used an algorithm based on deep learning and an evolutionary model. The performance of the proposed algorithm was better compared to other algorithms. Liu et al. [21] proposed the seasonal autoregressive integrated moving average (SARIMA) algorithm for predicting WS in Scotland. The results showed that this algorithm’s performance was better than the gated recurrent unit (GRU) and LSTM networks. Wang et al. [22] used LSTM networks to analyze WS’s long-term and short-term memory characteristics. The results showed the proper performance of this model. Liang et al. [23] used Bi-LSTM for WS prediction. The results confirmed the better performance of this algorithm compared to other algorithms. Xie et al. [24] used a multivariable LSTM network model to predict short-term WS. The results proved the superiority of this model compared to other models. Fu et al. [25] proposed improved hybrid differential evolution-Harris hawks optimization (IHDEHHO) and kernel extreme learning machine (KELM) for multi-step WS predicting. The good performance results confirmed the proposed algorithm.
Another WS prediction approach is the ensemble prediction approach. Ibrahim et al. [26] predicted WS using an ensemble prediction method based on deep learning and optimization algorithms. The results showed that the adaptive dynamic-particle swarm optimization-guided whale optimization algorithm (AD-PSO-Guided WOA) performed better than other algorithms. Qu et al. [27] proposed a new ensemble predicting approach based on optimization algorithms for short-term WS predicting. The results showed that the proposed approach is highly accurate. Liu et al. [28] used an ensemble prediction system based on ML algorithms to predict WS. Karthikeyan et al. [29] predicted WS using ensemble learning. The results showed that ensemble learning effectively predicts WS. Sloughter et al. [30] predicted WS using Bayesian model averaging (BMA). The use of this approach provided accurate predictions.
A review of the literature shows that there have been no reports on the use of the SEML method in modeling and predicting WS. Sometimes different algorithms may perform poorly in modeling and predicting WS data. The purpose of stacking is to exploit the potential of base algorithms and construct a more reliable framework than the base algorithms. Therefore, in this study, a novel framework was proposed using the SEML method for the first time. The structure of this method has two levels, which include base algorithms and meta-algorithm. For this purpose, eleven ML algorithms, including ANN, MLR, MNLR, GRNN, MARS, M5, GBR, RBFNN, LSBoost, LSSVM-GS, and LSSVM-HHO, in level 1 of the SEML method were used to model WS at sixteen stations in Iran (i.e., Quchan, Sarakhs, Sabzevar, Golmakan, Mashhad, Neyshabur, Torbat Heydarieh, Kashmar, Gonabad, Tabas, Ferdows, Qaen, Birjand, Nehbandan, Boshruyeh, and Khor Birjand). Level 2 of the proposed method applied the results stacked in level 1 to the LSBoost. This dramatically increased the accuracy of WS modeling. Thus, a novel framework using a set of ML algorithms based on the SEML method for modeling and predicting WS and other meteorological parameters is proposed.

2. Materials and Methods

2.1. Present Work Steps

This paper proposes a novel framework for modeling WS. The SEML method combines a wide range of ML algorithms, each of which performs well in a particular field and performs better than any single model in the set. The SEML method works well when several ML algorithms with different properties, structures, and assumptions are used as base algorithms. Base algorithms must be complex and diverse to have different benefits. This novel method allows for the simultaneous use of several algorithms and increases modeling accuracy by using different advantages of algorithms. The pseudo-code of the proposed framework based on the SEML method is described in Algorithm 1. The SEML method consists of two levels:
  • Level 1 (base algorithms): All algorithms perform their operations separately at this level. The outputs of this level’s algorithms are used as input into the meta-algorithm.
  • Level 2 (meta-algorithm): At this level, the meta-algorithm produces high-precision outputs by combining the results of the level 1 algorithms.
Algorithm1. Pseudo-code of the proposed framework based on the SEML method.
  • 1: Begin
  • 2: For each station, do:
  •         %% Base modeling
  •                  3:         Load inputs and outputs data for the current station;
  •                  4:         Divide data into training and testing periods using random calibration;
  •                  5:         Train the ANN, MLR, MNLR, GRNN, MARS, M5, GBR, RBFNN, LSBoost, LSSVM-GS, and LSSVM-HHO using training data;
  •                  6:         Test the ANN, MLR, MNLR, GRNN, MARS, M5, GBR, RBFNN, LSBoost, LSSVM-GS, and LSSVM-HHO using testing data;
  •                  7:         Evaluate the performance of the algorithms;
  •                  8:         Save the output of the ANN, MLR, MNLR, GRNN, MARS, M5, GBR, RBFNN, LSBoost, LSSVM-GS, and LSSVM-HHO in one matrix.
  •         %% Meta modeling
  •                  9:         Consider the output of the ANN, MLR, MNLR, GRNN, MARS, M5, GBR, RBFNN, LSBoost, LSSVM-GS, and LSSVM-HHO as input data;
  •                  10:         Consider the WS in the current station as output;
  •                  11:         Train the LSBoost using the considered inputs and output data;
  •                  12:         Test the LSBoost using the considered inputs and output data;
  •                  13:         Evaluate the performance of the LSBoost;
  •                  14:         Return the modeled WS to the current station.
  • 15: End
According to Figure 1, the following steps were performed to model WS using the SEML method:
  • Level 1: WS was modeled using eleven ML algorithms, including ANN, MLR, MNLR, GRNN, MARS, M5, GBR, RBFNN, LSBoost, LSSVM-GS, and LSSVM-HHO, at sixteen stations.
  • Level 2: Due to the simple structure and high speed of the LSBoost, this algorithm was used as a meta-algorithm to combine the results of the base algorithms.

2.2. Artificial Neural Network (ANN)

ANN is one of the most popular ML algorithms with a short computational time. It is capable of nonlinear modeling functions [31]. ANN is inspired by a process similar to the human brain to solve various problems. ANN is a data-driven algorithm used to develop the relationship between input and output variables [32]. The ANN comprises different layers (i.e., input, hidden, and output), each containing several neurons (Figure 2). Recently, ANN has been used to model various parameters. In ANN, the relationship between input (x) and output (Y) is as follows:
Y = f   ( w 1 x 1 + w 2 x 2 + + w n x n + b )
where g is the stimulation function; b is the bias; w n is the weight of link.

2.3. Multiple Linear Regression (MLR)

MLR is a statistical method used to model and analyze data [33]. MLR models output data based on a linear relationship [34]. MLR is an extension of ordinary least squares regression because it involves more than one explanatory variable. The MLR is an algorithm for examining the effect of a dependent variable by identifying the relationship of independent variables. The relationship between inputs and goals is assumed to be as follows:
Y = B 0 + B 1 x 1 + B 2 x 2 + + B n x n + ε
where Y is the dependent variable; x1, x2,…, xn are the independent variables; B1, B2,…, Bn are the coefficients of the independent variables; ε refers to the error term.

2.4. Multiple Nonlinear Regression (MNLR)

The MNLR, unlike the MLR, uses nonlinear regression analysis. In other words, this algorithm estimates the output data by creating a nonlinear combination [35]. The MNLR has more flexibility in modeling high-dimensional data due to the nonlinear relations. The relationship between input and output values in MNLR is as follows:
Y =   B 0 + B 1 x 1 + B 2 x 2 + + B n x n + B n + 1 x 1 2 + B n + 2 x 2 2 + + B 2 n x n 2 + ε

2.5. General Regression Neural Network (GRNN)

GRNN, obtained from neural network improvement, operates based on nonparametric regression. The GRNN uses the Gaussian kernel regression statistical method to solve various problems [36]. This algorithm does not require an iterative method for data modeling. This ability approximates any desired function between input and output data. The GRNN consists of four layers: input, pattern, summation, and output [37]. The relationship between input (x) and output (Y) in this algorithm is as follows:
Y = i = 1 n Y i   exp ( ( x x i ) T ( x x i ) 2 σ 2 ) i = 1 n exp ( ( x x i ) T ( x x i ) 2 σ 2 )
where Yi is the weight of the ith neuron connection between the layers; σ is the smoothing factor.

2.6. Multivariate Adaptive Regression Splines (MARS)

The MARS, a nonlinear and nonparametric method, was first proposed by Friedman [38]. This algorithm, which has good performance in solving nonlinear problems with high dimensions, is used to model and predict various parameters. The MARS, which consists of basic functions (BFs), defines nonlinear relationships by determining the weight of input variables. Regression models are determined by combining BFs to estimate the outputs. In this algorithm, BFs depend on spline functions that divide the data set into smaller parts. The structure of this model can be described as follows:
BF i = max ( 0 ,   x c i )   or   max ( 0 ,   c i x )
Y = i = 1 n w i BF i ( x )
where x is an input; c is a node of the ith BF; Y is an output; wi is the weight of the ith BF; n is the number of BFs; i is the known counter of BF.

2.7. M5 Model Tree (M5)

The M5, a data mining algorithm, was first proposed by Quinlan [39]. The M5 operates in two phases. It first divides the input region into several subsets and uses a specific linear regression model for each subset. Then, a linear regression is established in each node. The structure of the M5 is like a tree growing from the root (first node) to the leaves [40]. Branches are made up of numbers from the initial node to ninety. The growth criterion of the M5 is standard deviation reduction [41]. The random forest estimate is according to specific weights. This algorithm has an acceptable performance due to the random algorithms in a tree formation. The prevailing relation in the M5 is the standard deviation reduction:
SDR = SD ( T ) i | T i | | T | SD ( T i )
where T is a subset of data at each node; Ti is the set of attribute values obtained by dividing T; SD is the standard deviation.

2.8. Gradient Boosted Regression (GBR)

GBR consists of decision trees used for regression and classification tasks [42]. The main structure of GBR is that new models are added to the collection to strengthen the previous model at each stage. This algorithm creates a high-precision prediction model by combining weak models [43]. The advantage of this algorithm is that it can manage lost data. Another advantage of GBR is that it offers several hyper-parameter tuning options, which increases the flexibility of this algorithm. The mathematical formulation of GBR is as follows:
Y = m = 1 M γ m h m ( x )
where h m ( x ) are the basis functions; γ m are the shrinkage parameters.

2.9. Radial Basis Function Neural Network (RBFNN)

The RBFNN is a feedforward network consisting of three main layers: an input layer, a hidden layer, and an output layer [44]. The input layer is made up of neurons that carry information to the next layer. The hidden layer processes the input information. The nodes in the hidden layer are composed of radial basis functions (RBFs). An RBF is a symmetric function that applies a nonlinear map to input data. The number of hidden layer nodes varies according to the problem under study. The RBFNN uses various hidden layer functions that greatly affect how the algorithm works [45]. The output (Y) of the RBFNN algorithm is as follows:
Y = i = 1 j w i φ i ( x )
where wi is the weight of neurons; φ i ( x ) is the activation function of the RBFNN.

2.10. Least Squares Boost (LSBoost)

The LSBoost, first introduced in 2001 [46], is a deep and group learning method that performs perfectly in modeling and predicting various problems [2]. This algorithm, one of the most advanced gradient amplification algorithms, was developed based on the least squares, absolute, and Huber loss function and is used to increase the accuracy and reduce the variance of the results compared to independent learner samples. This algorithm’s simple structure, high speed, and accuracy are its other advantages. LSBoost uses the following equation to estimate target values:
F M ( x ) = F 0 ( x ) + v * i = 1 n α m B m ( x )
where F 0 ( x ) is the aggregated prediction for input values ( x ) ; B1,…, Bm are standalone weak learners; α 1 . . α M are the weights of the weak learners; v * is the learning ratio.

2.11. Least Squares Support Vector Machine-Grid Search (LSSVM-GS)

The LSSVM was first proposed by Suykens et al. [47]. The LSSVM is an improved algorithm based on the SVM used for regression analysis, classification, and pattern recognition. This algorithm uses linear relationships instead of nonlinear relationships between inputs (x) and output (M) and significantly improves the accuracy and speed of the SVM. The regression function between input and output in this algorithm is as follows:
M = α 1 k ( x ,   x 1 ) + α 2 k ( x ,   x 2 ) + + α i k ( x ,   x i ) + b
where α 1 is the ith Lagrange multiplier; K is a kernel function; b is the bias. The parameters α i and b are computed as follows:
[ kernel 1 ¯ T 1 ¯ kernel + C 1 I ] [ b α ] = [ 0 M ]
where C is the regulation parameter. The parameters α, M, I, and 1 ¯ are computed as follows:
α = [ α 1 α n ] ,   1 ¯ = [ 1 1 ]   ,   M = [ M 1 M n ]   ,   I = diag ( 1 , 1 , .1 )
In this paper, the radial basis function was used as a kernel function, which is expressed as follows [48]:
k ( x ,   x i ) = exp ( x x i 2 2 σ 2 )
where σ is an effective parameter in the accuracy of the algorithm. The values of C and the core width (σ) have a large effect on the accuracy of the LSSVM. GS can be used to determine the values of these coefficients. GS takes all of the values of C and σ and builds a grid within a certain range. Figure 3 shows the structure of the LSSVM-GS.

2.12. Harris Hawks Optimization (HHO)

HHO is a population-based algorithm first proposed by Heidari et al. [49]. The performance of this algorithm is inspired by the cooperative behaviors and chasing styles of Harris’ hawks. In this behavior, the Harris’ hawks attack from different directions to surprise the prey. This algorithm has been used in various issues due to the fact of its flexible structure, accuracy, and high speed [50]. For more details, see Heidari et al. [49].

2.13. LSSVM-HHO

Since the C and σ parameters significantly impact how the LSSVM works, it is essential to determine the optimal values of these parameters. Therefore, this study used the HHO optimization algorithm to find the optimal value of the LSSVM parameters. The LSSVM-HHO flow chart is shown in Figure 4. The values, C and σ, are considered decision variables in the hybrid algorithm. In this algorithm, first, training and testing data are randomly selected. Then, the initial parameters of HHO and the initial population of LSSVM are initialized to search for the C and σ values. In the next step, first, the training data are used to obtain the optimal LSSVM. Finally, the testing data are used to evaluate the predictive ability of the LSSVM.

2.14. Performance Evaluation of the Algorithms

In the present study, various evaluation criteria, mean absolute error (MAE), root mean squared error (RMSE), relative root mean square error (RRMSE), and R were employed to evaluate the performance of the ML algorithms. The evaluation criteria are given as follows [51,52]:
MAE = 1 N i = 1 N | x i y i |
RMSE = ( i = 1 N ( x i y i ) 2 N )
RRMSE = ( i = 1 N ( x i y i ) 2 N ) / SD ( y )
R = i = 1 N ( y i y ¯ ) ( x i x ¯ ) i = 1 N ( y i y ¯ ) i = 1 N ( x i x ¯ )
where N is the number of data; x is the observed values; y is the estimated values; x ¯ is the mean observed values; y ¯ is the mean estimated values. The MAE, RMSE, and RRMSE are in the range of zero to infinite positive numbers. The smaller the MAE, RMSE, and RRMSE values, the better the algorithm’s performance. Moreover, the higher the R value, the closer the modeled values are to the actual values, and the best value is one.

2.15. Case Study and Data Sources

In this study, sixteen synoptic stations located in the east and northeast of Iran were used to model WS and prove the performance of the proposed framework. Nine stations, including Quchan, Sarakhs, Sabzevar, Golmakan, Mashhad, Neyshabur, Torbat Heydarieh, Kashmar, and Gonabad, are located in Khorasan Razavi Province, and seven stations, including Tabas, Ferdows, Qaen, Birjand, Nehbandan, Boshruyeh, and Khor Birjand, are located in South Khorasan Province. The geographical characteristics of the studied stations are stated in Table 1. This region has high potential for the construction of wind farms. Due to the presence of one of the most important wind power plants in Iran in Neyshabur, modeling and predicting WS in this region is very important. In addition, many wind power plants are being built in this area, proving the importance of studied stations. Figure 5 shows the position of the stations used.
The statistical specifications on the data from different stations are given in Table 1. According to Table 1, seven parameters, including minimum air temperature (Tmin), mean air temperature (Tmean), maximum air temperature (Tmax), minimum relative humidity (Hmin), mean relative humidity (Hmean), maximum relative humidity (Hmax), and sunny hours (SH), were used for WS modeling.
Figure 6 and Figure A1 (Appendix A) show the probability density function (PDF) plot of WS data at the different stations. These figures analyze the volatility of the WS data set in the studied stations. According to these plots, the occurrence probability of any value of WS is calculated. Any WS range that is more likely to occur is known as the prevailing WS of that station.

3. Results and Discussion

3.1. Modeling WS using the SEML Method (Level 1—Base Algorithms)

Because different algorithms have unique features, each of them models only part of the data with good accuracy and does not exhibit good accuracy in other parts of the data. Therefore, using the SEML method combines the capabilities of all algorithms, and the modeling of different parts of the data is performed better. Level 1 of the SEML method aimed to use eleven ML algorithms with different natures and structures and different categories for WS modeling. For this reason, neuron-based, kernel-based, tree-based, curve-based, regression-based, and hybrid algorithm-based algorithms were used in this paper. Table 2, Table A1, Table A2, Table A3, Table A4, Table A5, Table A6, Table A7, Table A8, Table A9, Table A10, Table A11 and Table A12 (Appendix B) state the values of the evaluation criteria for the different algorithms at sixteen stations. According to these tables, the correlation coefficients obtained by the algorithms at all stations ranged from −0.01 to 0.62. This meant the poor performance of ML algorithms in modeling and estimating WS. The large fluctuations of WS values in the studied stations made their modeling difficult. For this reason, the eleven algorithms used in level 1 of the SEML method did not perform well in estimating the target data.
Table 3 lists the average evaluation criteria obtained by the base algorithms (during the testing period) at sixteen stations. According to this table, the average value of MAE was in the range of 0.96–1.29, the average value of RMSE was in the range of 1.31–1.66, the average value of RRMSE was in the range of 0.84–1.10, and the average value of R was in the range of 0.12–0.51. The average results of WS modeling indicated the low accuracy of the different algorithms. Despite different advantages, the different shortcomings in the structure of the base algorithms led to low modeling accuracy. The proposed framework based on SEML method aimed to eliminate the defects of the base algorithms and increase the modeling accuracy. The SEML method, as an integrated system, combined the results obtained by the level 1 algorithms and applied them to a meta-algorithm. Combining the algorithms’ results meant using all of the features of the base algorithms in WS modeling.

3.2. Modeling WS using the SEML Method (Level 2—Meta-Algorithm)

Due to the poor performance of the ML algorithms in WS modeling, the SEML method was used to increase the accuracy of the algorithms. For this purpose, the LSBoost was used as a meta-algorithm at level 2 of the SEML method. The reason for choosing this algorithm as a meta-algorithm was its simple structure, high accuracy, high speed, and easy execution. At level 2, the outputs obtained by the eleven algorithms were used as the stacked data as the meta-algorithm inputs. Table 4 shows the results of WS modeling by the LSBoost at sixteen stations. As the results show, the average value of MAE was in the range of 0.26–0.71, the average value of RMSE was in the range of 0.44–1.29, the average value of RRMSE was in the range of 0.46–0.60, and the average value of R was in the range of 0.80–0.89. The SEML method increased the WS modeling accuracy by more than 43%. In addition, the values of the other evaluation criteria obtained by the meta-algorithm at the different stations improved the WS data modeling’s accuracy. According to Table 4, the average values of MAE, RMSE, RRMSE, and R in modeling WS at sixteen stations by the meta-algorithm (LSBoost) were 0.43, 0.78, 0.51, and 0.86, respectively. A comparison of the average evaluation criteria obtained by the LSBoost in Table 3 and Table 4 showed the effect of the SEML method on improving the accuracy of WS modeling. Simultaneous use of the advantages of different algorithms at level 2 of the SEML method by LSBoost led to a significant increase in modeling accuracy. In addition, using different categories in the base algorithms led to using the advantages and features of the different algorithms in this framework. This eliminated the shortcomings of the algorithms in the proposed framework. According to the modeling results with LSBoost, the highest modeling accuracies were obtained at Quchan station (MAE = 0.41, RMSE = 0.71, RRMSE = 0.46, and R = 0.89) and Neyshabur station (MAE = 0.26, RMSE = 0.44, RRMSE = 0.49, and R = 0.87).
Figure 7, Figure 8, Figure A2 and Figure A3 (Appendix C) show the results of the scatter plot and time series of the observed data and the data modeled by the LSBoost. It is clear from these figures that the proposed novel framework successfully performed WS modeling. In addition, the use of the SEML method in WS modeling significantly increased the accuracy of the algorithms, so that the values of the evaluation criteria in level 2 were improved compared to the base algorithms at all stations. According to the scatterplots, the difference between the values modeled by the meta-algorithm and the observed values was very small. Since most of the modeled and observed values were collected around the semiconductor line, the accuracy of the LSBoost in WS modeling was high. Moreover, according to the time series diagrams, the data model at level 2 of the SEML method was largely matched to the observed data. Accurate estimations of the maximum and minimum WS values by meta-algorithm show the simultaneous use of the advantages of base algorithms and their application in level 2 of the SEML method.
Figure 9 and Figure A4 (Appendix D) show the quantile–quantile plot (Q–Q plot) at different stations. In these plots, the probability distributions of the observed data were plotted against the probability distributions of the data modeled by the meta-algorithm (LSBoost). The more points collected around line 45°, the more equal the probabilistic distributions of the observed and modeled data. The accumulation of points around the 45° line also showed a slight difference between the observed and modeled data. According to the descriptions provided during the test period, the observed and modeled data had an equal distribution at most stations. This means that the LSBoost in the structure of the SEML method modeled the WS data with high accuracy.

4. Conclusions

Accurate and reliable modeling and predicting WS is critical to planning and managing power generation. WS modeling was performed at sixteen stations in Iran at two levels to achieve this goal. In level 1 of this method, eleven ML algorithms, including ANN, MLR, MNLR, GRNN, MARS, M5, GBR, RBFNN, LSBoost, LSSVM-GS, and LSSVM-HHO, in six categories (neuron based, kernel based, tree based, curve based, regression based, and hybrid algorithm based) were used as base algorithms. In level 2, the outputs of the base algorithms were applied as input to the meta-algorithm (LSBoost). Thus, the final results of WS modeling were obtained. Applying the results of the base algorithms to the LSBoost at level 2 of the SEML method increased the modeling accuracy by >43%. Improving the evaluation criteria at all stations proved the strength of the novel framework. The average values of MAE, RMSE, RRMSE, and R in modeling WS at sixteen stations using the meta-algorithm (LSBoost) were 0.43, 0.78, 0.51, and 0.86, respectively. According to the results of this study, the proposed framework has high potential for analysis of other engineering issues and meteorological parameters. This framework can improve the performance of ML algorithms in modeling and predicting WS in areas where ML algorithms perform poorly. Moreover, using this novel framework, WS can be predicted in different basins and the optimal use of WS and electricity generation can be planned. The performance of the proposed framework can be improved by using appropriate algorithms and various techniques such as wavelet transform. As a result, we achieved the desired goal of providing a novel framework based on the SEML method.

Author Contributions

Conceptualization, S.F.; Data curation, A.M.-B., M.K. and M.V.A.; Formal analysis, A.M.-B., M.K., M.V.A. and S.F.; Investigation, A.M.-B., M.K. and S.F.; Methodology, A.M.-B., M.K. and M.V.A.; Project administration, S.F.; Resources, S.F. and M.V.A.; Software, A.M.-B., M.K. and M.V.A.; Supervision, S.F.; Validation, A.M.-B., M.K. and M.V.A.; Visualization, A.M.-B., M.K. and M.V.A.; Writing—original draft, A.M.-B. and M.K.; Writing—review and editing, S.F. and M.V.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research was not supported through any funds.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All data generated or used during the study are applicable if requested.

Conflicts of Interest

The authors declare that they have no conflict of interest.

Appendix A

Figure A1. PDF plots of observed values of WS at different stations.
Figure A1. PDF plots of observed values of WS at different stations.
Atmosphere 13 00758 g0a1aAtmosphere 13 00758 g0a1b

Appendix B

Table A1. Results of modeling WS with the base algorithms at the Sarakhs station.
Table A1. Results of modeling WS with the base algorithms at the Sarakhs station.
TrainingTesting
MAERMSERRMSERMAERMSERRMSER
ANN1.101.500.890.461.111.520.910.42
MLR1.181.580.940.341.171.580.950.32
MNLR1.161.560.930.371.151.560.940.35
GRNN0.861.210.720.741.131.560.930.36
MARS1.141.540.910.411.131.540.920.40
M51.131.530.910.411.131.560.930.37
GBR1.031.390.830.591.111.520.910.42
RBFNN0.000.000.001.001.431.811.08−0.01
LSBoost1.061.410.840.571.121.540.920.40
LSSVM-GS0.951.280.760.681.111.520.910.42
LSSVM-HHO1.031.390.820.591.111.520.910.42
Table A2. Results of modeling WS with the base algorithms at the Sabzevar station.
Table A2. Results of modeling WS with the base algorithms at the Sabzevar station.
TrainingTesting
MAERMSERRMSERMAERMSERRMSER
ANN0.921.230.860.510.981.290.880.47
MLR0.931.250.870.490.981.300.890.46
MNLR0.931.240.860.500.981.290.880.47
GRNN0.690.970.680.750.981.300.890.47
MARS0.921.230.860.510.981.290.880.47
M50.911.220.850.530.991.300.890.46
GBR0.781.050.730.700.971.280.870.49
RBFNN0.000.000.001.001.711.951.330.00
LSBoost0.891.180.830.570.971.290.880.48
LSSVM-GS0.781.040.730.700.971.280.870.49
LSSVM-HHO0.781.050.730.700.971.280.870.49
Table A3. Results of modeling WS with the base algorithms in the Golmakan station.
Table A3. Results of modeling WS with the base algorithms in the Golmakan station.
TrainingTesting
MAERMSERRMSERMAERMSERRMSER
ANN1.351.920.970.251.452.070.980.20
MLR1.381.960.990.161.462.080.990.17
MNLR1.371.950.980.191.452.070.980.18
GRNN0.991.440.730.751.482.121.010.11
MARS1.371.940.980.211.452.070.980.19
M51.351.920.970.251.472.090.990.16
GBR1.291.840.930.431.452.070.980.20
RBFNN0.000.000.001.001.482.111.000.01
LSBoost1.331.880.950.351.452.070.980.18
LSSVM-GS1.151.630.820.691.452.070.980.19
LSSVM-HHO1.291.840.930.431.452.070.980.20
Table A4. Results of modeling WS with the base algorithms at the Mashhad station.
Table A4. Results of modeling WS with the base algorithms at the Mashhad station.
TrainingTesting
MAERMSERRMSERMAERMSERRMSER
ANN0.771.020.870.490.801.040.890.46
MLR0.791.050.900.430.811.060.910.42
MNLR0.791.050.900.440.811.060.900.43
GRNN0.580.790.670.770.801.070.910.42
MARS0.791.040.890.450.811.050.900.43
M50.781.040.890.450.811.070.910.41
GBR0.720.950.820.590.801.050.890.45
RBFNN0.000.000.001.002.202.492.130.00
LSBoost0.761.000.860.520.801.050.900.44
LSSVM-GS0.660.870.750.690.801.050.900.44
LSSVM-HHO0.720.950.820.590.801.050.890.45
Table A5. Results of modeling WS with the base algorithms at the Neyshabur station.
Table A5. Results of modeling WS with the base algorithms at the Neyshabur station.
TrainingTesting
MAERMSERRMSERMAERMSERRMSER
ANN0.590.790.860.510.590.800.880.48
MLR0.610.810.890.460.610.820.900.44
MNLR0.600.800.880.470.610.810.890.45
GRNN0.450.630.700.740.610.820.900.44
MARS0.590.790.870.500.590.800.880.48
M50.590.790.870.490.600.810.890.46
GBR0.510.690.760.670.590.790.870.49
RBFNN0.000.000.001.000.710.911.000.00
LSBoost0.570.760.830.560.600.800.880.47
LSSVM-GS0.500.670.740.700.590.790.870.49
LSSVM-HHO0.510.690.760.670.590.790.870.49
Table A6. Results of modeling WS with the base algorithms at the Kashmar station.
Table A6. Results of modeling WS with the base algorithms at the Kashmar station.
TrainingTesting
MAERMSERRMSERMAERMSERRMSER
ANN0.751.010.910.420.751.010.900.45
MLR0.771.040.930.370.761.030.910.41
MNLR0.771.030.920.380.761.030.910.42
GRNN0.580.810.720.720.771.040.920.39
MARS0.751.020.910.410.751.020.900.44
M50.751.020.910.410.751.020.900.44
GBR0.680.930.830.580.751.010.900.45
RBFNN0.000.000.001.000.841.110.980.22
LSBoost0.710.950.850.540.751.020.900.43
LSSVM-GS0.640.870.780.670.751.020.900.44
LSSVM-HHO0.640.870.780.670.751.020.900.44
Table A7. Results of modeling WS with the base algorithms at the Ferdows station.
Table A7. Results of modeling WS with the base algorithms at the Ferdows station.
TrainingTesting
MAERMSERRMSERMAERMSERRMSER
ANN0.771.030.880.470.801.080.900.44
MLR0.791.070.910.420.811.090.910.41
MNLR0.791.060.900.430.811.090.910.42
GRNN0.570.800.680.760.811.100.920.40
MARS0.791.050.900.440.801.080.900.43
M50.781.050.890.460.821.110.920.39
GBR0.720.970.820.580.801.070.900.44
RBFNN0.000.000.001.000.941.211.010.03
LSBoost0.730.970.830.580.801.080.900.43
LSSVM-GS0.660.890.760.690.801.070.900.44
LSSVM-HHO0.720.970.820.580.801.070.900.44
Table A8. Results of modeling WS with the base algorithms at the Qaen station.
Table A8. Results of modeling WS with the base algorithms at the Qaen station.
TrainingTesting
MAERMSERRMSERMAERMSERRMSER
ANN0.901.210.800.610.941.260.830.56
MLR0.961.280.840.540.991.320.870.50
MNLR0.941.260.830.560.971.300.850.53
GRNN0.700.980.640.780.961.300.850.53
MARS0.921.230.810.590.941.270.830.55
M50.921.240.820.580.951.270.840.55
GBR0.831.120.730.690.941.260.830.56
RBFNN0.000.000.001.001.361.741.140.00
LSBoost0.861.150.760.660.951.270.840.55
LSSVM-GS0.771.040.680.740.941.270.830.56
LSSVM-HHO0.831.120.730.690.941.260.830.56
Table A9. Results of modeling WS with the base algorithms at the Birjand station.
Table A9. Results of modeling WS with the base algorithms at the Birjand station.
TrainingTesting
MAERMSERRMSERMAERMSERRMSER
ANN0.871.170.870.490.901.200.880.47
MLR0.921.230.910.410.921.230.900.43
MNLR0.901.210.900.440.911.210.890.46
GRNN0.680.960.710.720.911.210.890.45
MARS0.891.190.880.470.901.200.880.47
M50.881.190.880.480.901.210.890.46
GBR0.751.000.740.690.891.190.870.49
RBFNN0.000.000.001.001.031.361.000.02
LSBoost0.771.020.760.680.901.190.880.48
LSSVM-GS0.730.980.730.720.891.190.870.49
LSSVM-HHO0.751.000.740.690.891.190.870.49
Table A10. Results of modeling WS with the base algorithms at the Nehbandan station.
Table A10. Results of modeling WS with the base algorithms at the Nehbandan station.
TrainingTesting
MAERMSERRMSERMAERMSERRMSER
ANN1.231.720.770.641.281.800.790.62
MLR1.351.850.830.561.361.890.830.56
MNLR1.291.780.790.611.301.820.800.60
GRNN1.011.490.670.751.291.810.790.61
MARS1.271.760.780.621.281.800.790.62
M51.251.750.780.631.291.820.800.60
GBR1.171.630.730.691.271.780.780.62
RBFNN0.000.000.001.001.542.210.970.36
LSBoost1.221.680.750.661.291.810.790.61
LSSVM-GS1.071.500.670.751.271.790.780.62
LSSVM-HHO1.171.630.730.691.271.780.780.62
Table A11. Results of modeling WS with the base algorithms in the Boshruyeh station.
Table A11. Results of modeling WS with the base algorithms in the Boshruyeh station.
TrainingTesting
MAERMSERRMSERMAERMSERRMSER
ANN0.781.080.900.430.771.080.900.43
MLR0.821.120.940.350.811.100.920.38
MNLR0.821.120.940.350.801.100.920.39
GRNN0.600.860.720.730.801.110.930.37
MARS0.801.110.930.380.791.090.910.41
M50.801.100.920.390.801.090.920.40
GBR0.680.940.780.660.781.070.900.43
RBFNN0.000.000.001.000.931.191.000.00
LSBoost0.751.020.850.540.791.080.910.41
LSSVM-GS0.680.930.780.670.781.070.900.43
LSSVM-HHO0.680.930.780.670.781.070.900.43
Table A12. Results of modeling WS with base algorithms in the Khor Birjand station.
Table A12. Results of modeling WS with base algorithms in the Khor Birjand station.
TrainingTesting
MAERMSERRMSERMAERMSERRMSER
ANN1.692.230.860.511.702.230.860.50
MLR1.742.290.880.471.752.290.890.46
MNLR1.712.260.870.501.722.260.870.48
GRNN1.291.780.680.751.742.280.880.47
MARS1.692.230.860.521.712.240.870.50
M51.682.220.850.531.742.280.880.48
GBR1.592.100.800.601.712.240.870.50
RBFNN0.000.000.001.001.892.490.960.31
LSBoost1.582.070.800.611.712.240.870.49
LSSVM-GS1.441.900.730.701.722.240.870.50
LSSVM-HHO1.592.100.800.601.712.240.870.50

Appendix C

Figure A2. Scattered plots of observed and modeled values of WS by meta-algorithm at different stations.
Figure A2. Scattered plots of observed and modeled values of WS by meta-algorithm at different stations.
Atmosphere 13 00758 g0a2aAtmosphere 13 00758 g0a2bAtmosphere 13 00758 g0a2c
Figure A3. Time series of observed and modeled values of WS using the meta-algorithm at different stations.
Figure A3. Time series of observed and modeled values of WS using the meta-algorithm at different stations.
Atmosphere 13 00758 g0a3aAtmosphere 13 00758 g0a3b

Appendix D

Figure A4. Q–Q plots of the observed and modeled values of WS using the meta-algorithm at different stations.
Figure A4. Q–Q plots of the observed and modeled values of WS using the meta-algorithm at different stations.
Atmosphere 13 00758 g0a4aAtmosphere 13 00758 g0a4bAtmosphere 13 00758 g0a4cAtmosphere 13 00758 g0a4d

References

  1. Singh, U.; Rizwan, M.; Alaraj, M.; Alsaidan, I. A Machine Learning-Based Gradient Boosting Regression Approach for Wind Power Production Forecasting: A Step towards Smart Grid Environments. Energies 2021, 14, 5196. [Google Scholar] [CrossRef]
  2. Jung, C.; Schindler, D. Integration of small-scale surface properties in a new high resolution global wind speed model. Energy Convers. Manag. 2020, 210, 112733. [Google Scholar] [CrossRef]
  3. Busico, G.; Kazakis, N.; Cuoco, E.; Colombani, N.; Tedesco, D.; Voudouris, K.; Mastrocicco, M. A novel hybrid method of specific vulnerability to anthropogenic pollution using multivariate statistical and regression analyses. Water Res. 2020, 171, 115386. [Google Scholar] [CrossRef] [PubMed]
  4. Farzin, S.; Chianeh, F.N.; Anaraki, M.V.; Mahmoudian, F. Introducing a framework for modeling of drug electrochemical removal from wastewater based on data mining algorithms, scatter interpolation method, and multi criteria decision analysis (DID). J. Clean. Prod. 2020, 266, 122075. [Google Scholar] [CrossRef]
  5. Kadkhodazadeh, M.; Farzin, S. A Novel LSSVM Model Integrated with GBO Algorithm to Assessment of Water Quality Parameters. Water Resour. Manag. 2021, 35, 3939–3968. [Google Scholar] [CrossRef]
  6. Jha, M.K.; Shekhar, A.; Jenifer, M.A. Assessing groundwater quality for drinking water supply using hybrid fuzzy-GIS-based water quality index. Water Res. 2020, 179, 115867. [Google Scholar] [CrossRef]
  7. Ghorbani, M.A.; Jabehdar, M.A.; Yaseen, Z.M.; Inyurt, S. Solving the pan evaporation process complexity using the development of multiple mode of neurocomputing models. Theor. Appl. Climatol. 2021, 145, 1521–1539. [Google Scholar] [CrossRef]
  8. Zhou, J.; Shi, J.; Li, G. Fine tuning support vector machines for short-term wind speed forecasting. Energy Convers. Manag. 2011, 52, 1990–1998. [Google Scholar] [CrossRef]
  9. Panapakidis, I.P.; Dagoumas, A.S. Day-ahead electricity price forecasting via the application of artificial neural network based models. Appl. Energy 2016, 172, 132–151. [Google Scholar] [CrossRef]
  10. Fogno Fotso, H.R.; Aloyem Kazé, C.V.; Djuidje Kenmoé, G.A. A novel hybrid model based on weather variables relationships improving applied for wind speed forecasting. Int. J. Energy. Environ. Eng. 2022, 13, 43–56. [Google Scholar] [CrossRef]
  11. Elsisi, M.; Tran, M.Q.; Mahmoud, K.; Lehtonen, M.; Darwish, M.M.F. Robust design of ANFIS-based blade pitch controller for wind energy conversion systems against wind speed fluctuations. IEEE Access 2021, 9, 37894–37904. [Google Scholar] [CrossRef]
  12. Huang, X.; Wang, J.; Huang, B. Two novel hybrid linear and nonlinear models for wind speed forecasting. Energy Convers. Manag. 2021, 238, 114162. [Google Scholar] [CrossRef]
  13. Wang, J.; Li, H.; Wang, Y.; Lu, H. A hesitant fuzzy wind speed forecasting system with novel defuzzification method and multi-objective optimization algorithm. Expert Syst. Appl. 2021, 168, 114364. [Google Scholar] [CrossRef]
  14. Chen, X.J.; Zhao, J.; Jia, X.Z.; Li, Z.L. Multi-step wind speed forecast based on sample clustering and an optimized hybrid system. Renew Energy 2021, 165, 595–611. [Google Scholar] [CrossRef]
  15. Du, P.; Wang, J.; Yang, W.; Niu, T. A novel hybrid model for short-term wind power forecasting. Appl. Soft Comput. 2019, 80, 93–106. [Google Scholar] [CrossRef]
  16. Cheng, Z.; Wang, J. A new combined model based on multi-objective salp swarm optimization for wind speed forecasting. Appl. Soft Comput. 2020, 92, 106294. [Google Scholar] [CrossRef]
  17. Chen, Y.; Dong, Z.; Wang, Y.; Su, J.; Han, Z.; Zhou, D.; Zhang, K.; Zhao, Y.; Bao, Y. Short-term wind speed predicting framework based on EEMD-GA-LSTM method under large scaled wind history. Energy Convers. Manag. 2021, 227, 113559. [Google Scholar] [CrossRef]
  18. Liu, H.; Duan, Z.; Chen, C.; Wu, H. A novel two-stage deep learning wind speed forecasting method with adaptive multiple error corrections and bivariate Dirichlet process mixture model. Energy Convers. Manag. 2019, 199, 111975. [Google Scholar] [CrossRef]
  19. Liu, Y.; Qin, H.; Zhang, Z.; Pei, S.; Jiang, Z.; Feng, Z.; Zhou, J. Probabilistic spatiotemporal wind speed forecasting based on a variational Bayesian deep learning model. Appl. Energy 2020, 260, 114259. [Google Scholar] [CrossRef]
  20. Neshat, M.; Nezhad, M.M.; Abbasnejad, E.; Mirjalili, S.; Tjernberg, L.B.; Astiaso Garcia, D.; Alexander, B.; Wagner, M. A deep learning-based evolutionary model for short-term wind speed forecasting: A case study of the Lillgrund offshore wind farm. Energy Convers. Manag. 2021, 236, 114002. [Google Scholar] [CrossRef]
  21. Liu, X.; Lin, Z.; Feng, Z. Short-term offshore wind speed forecast by seasonal ARIMA—A comparison against GRU and LSTM. Energy 2021, 227, 120492. [Google Scholar] [CrossRef]
  22. Wang, J.; Li, Y. Multi-step ahead wind speed prediction based on optimal feature extraction, long short term memory neural network and error correction strategy. Appl. Energy 2018, 230, 429–443. [Google Scholar] [CrossRef]
  23. Liang, T.; Zhao, Q.; Lv, Q.; Sun, H. A novel wind speed prediction strategy based on Bi-LSTM, MOOFADA and transfer learning for centralized control centers. Energy 2021, 230, 120904. [Google Scholar] [CrossRef]
  24. Xie, A.; Yang, H.; Chen, J.; Sheng, L.; Zhang, Q.; Kumar Jha, S.; Zhang, X.; Zhang, L.; Patel, N. A Short-Term Wind Speed Forecasting Model Based on a Multi-Variable Long Short-Term Memory Network. Atmosphere 2021, 12, 651. [Google Scholar] [CrossRef]
  25. Fu, W.; Zhang, K.; Wang, K.; Wen, B.; Fang, P.; Zou, F. A hybrid approach for multi-step wind speed forecasting based on two-layer decomposition, improved hybrid DE-HHO optimization and KELM. Renew Energy 2021, 164, 211–229. [Google Scholar] [CrossRef]
  26. Ibrahim, A.; Mirjalili, S.; El-Said, M.; Ghoneim, S.S.M.; Al-Harthi, M.M.; Ibrahim, T.F.; El-Kenawy, E.S.M. Wind Speed Ensemble Forecasting Based on Deep Learning Using Adaptive Dynamic Optimization Algorithm. IEEE Access 2021, 9, 125787–125804. [Google Scholar] [CrossRef]
  27. Qu, Z.; Zhang, K.; Mao, W.; Wang, J.; Liu, C.; Zhang, W. Research and application of ensemble forecasting based on a novel multi-objective optimization algorithm for wind-speed forecasting. Energy Convers. Manag. 2017, 154, 440–454. [Google Scholar] [CrossRef]
  28. Liu, Z.; Jiang, P.; Wang, J.; Zhang, L. Ensemble forecasting system for short-term wind speed forecasting based on optimal sub-model selection and multi-objective version of mayfly optimization algorithm. Expert Syst. Appl. 2021, 177, 114974. [Google Scholar] [CrossRef]
  29. Karthikeyan, M.; Rengaraj, R. Short-term wind speed forecasting using ensemble learning. In Proceedings of the 7th International Conference on Electrical Energy Systems (ICEES), Chennai, India, 11–13 February 2021; pp. 502–506. [Google Scholar] [CrossRef]
  30. Sloughter, J.M.L.; Gneiting, T.; Raftery, A.E. Probabilistic Wind Speed Forecasting Using Ensembles and Bayesian Model Averaging. J. Am. Stat. Assoc. 2012, 105, 25–35. [Google Scholar] [CrossRef] [Green Version]
  31. Valikhan Anaraki, M.; Mahmoudian, F.; Nabizadeh Chianeh, F.; Farzin, S. Dye Pollutant Removal from Synthetic Wastewater: A New Modeling and Predicting Approach Based on Experimental Data Analysis, Kriging Interpolation Method, and Computational Intelligence Techniques. J. Environ. Inform. 2022. [CrossRef]
  32. Tikhamarine, Y.; Malik, A.; Souag-Gamane, D.; Kisi, O. Artificial intelligence models versus empirical equations for modeling monthly reference evapotranspiration. Environ. Sci. Pollut. Res. 2020, 27, 30001–30019. [Google Scholar] [CrossRef] [PubMed]
  33. Lesar, T.T.; Filipčić, A. The Hourly Simulation of PM2.5 Particle Concentrations Using the Multiple Linear Regression (MLR) Model for Sea Breeze in Split, Croatia. Water Air Soil Pollut. 2021, 232, 261. [Google Scholar] [CrossRef]
  34. Kadkhodazadeh, M.; Anaraki, M.V.; Morshed-Bozorgdel, A.; Farzin, S. A New Methodology for Reference Evapotranspiration Prediction and Uncertainty Analysis under Climate Change Conditions Based on Machine Learning, Multi Criteria Decision Making and Monte Carlo Methods. Sustainability 2022, 14, 2601. [Google Scholar] [CrossRef]
  35. Huangfu, W.; Wu, W.; Zhou, X.; Lin, Z.; Zhang, G.; Chen, R.; Song, Y.; Lang, T.; Qin, Y.; Ou, P.; et al. Landslide Geo-Hazard Risk Mapping Using Logistic Regression Modeling in Guixi, Jiangxi, China. Sustainability 2021, 13, 4830. [Google Scholar] [CrossRef]
  36. Salehi, M.; Farhadi, S.; Moieni, A.; Safaie, N.; Hesami, M. A hybrid model based on general regression neural network and fruit fly optimization algorithm for forecasting and optimizing paclitaxel biosynthesis in Corylus avellana cell culture. Plant Methods 2021, 17, 13. [Google Scholar] [CrossRef] [PubMed]
  37. Sridharan, M. Generalized Regression Neural Network Model Based Estimation of Global Solar Energy Using Meteorological Parameters. Ann. Data Sci. 2021. [CrossRef]
  38. Friedman, J.H. Multivariate Adaptive Regression Splines. Ann. Stat. 1991, 19, 1–67. [Google Scholar] [CrossRef]
  39. Quinlan, J.R.; Quinlan, J.R. Learning with Continuous Classes. In Proceedings of the 5th Australian Joint Conference on Artificial Intelligence, Hobart, Austrilia, 16–18 November 1992; World Scientific: Singapore, 1992; pp. 343–348. [Google Scholar] [CrossRef]
  40. Emamifar, S.; Rahimikhoob, A.; Noroozi, A.A. An Evaluation of M5 Model Tree vs. Artificial Neural Network for Estimating Mean Air Temperature as Based on Land Surface Temperature Data by MODIS-Terra Sensor. Iran. J. Soil Water Res. 2014, 45, 423–433. [Google Scholar] [CrossRef]
  41. Bharti, B.; Pandey, A.; Tripathi, S.K.; Kumar, D. Modelling of runoff and sediment yield using ANN, LS-SVR, REPTree and M5 models. Hydrol. Res. 2017, 48, 1489–1507. [Google Scholar] [CrossRef]
  42. Arumugam, P.; Chemura, A.; Schauberger, B.; Gornott, C. Remote Sensing Based Yield Estimation of Rice (Oryza Sativa L.) Using Gradient Boosted Regression in India. Remote Sens. 2021, 13, 2379. [Google Scholar] [CrossRef]
  43. Wen, H.T.; Lu, J.H.; Phuc, M.X. Applying Artificial Intelligence to Predict the Composition of Syngas Using Rice Husks: A Comparison of Artificial Neural Networks and Gradient Boosting Regression. Energies 2021, 14, 2932. [Google Scholar] [CrossRef]
  44. Pazouki, G.; Golafshani, E.M.; Behnood, A. Predicting the compressive strength of self-compacting concrete containing Class F fly ash using metaheuristic radial basis function neural network. Struct. Concr. 2021, 23, 1191–1213. [Google Scholar] [CrossRef]
  45. Ojo, S.; Imoize, A.; Alienyi, D. Radial basis function neural network path loss prediction model for LTE networks in multitransmitter signal propagation environments. Int. J. Commun. Syst. 2021, 34, e4680. [Google Scholar] [CrossRef]
  46. Friedman, J.H. Greedy function approximation: A gradient boosting machine. Ann. Statist. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
  47. Suykens, J.A.K.; Van Gestel, T.; De Brabanter, J.; De Moor, B.; Vandewalle, J. Least Squares Support Vector Machines; World Scientific: Singapore, 2002. [Google Scholar] [CrossRef]
  48. Farzin, S.; Valikhan Anaraki, M. Modeling and predicting suspended sediment load under climate change conditions: A new hybridization strategy. J. Water Clim. Chang. 2021, 12, 2422–2443. [Google Scholar] [CrossRef]
  49. Heidari, A.A.; Mirjalili, S.; Faris, H.; Aljarah, I.; Mafarja, M.; Chen, H. Harris hawks optimization: Algorithm and applications. Future Gener. Comput. Syst. 2019, 97, 849–872. [Google Scholar] [CrossRef]
  50. Houssein, E.H.; Hosney, M.E.; Oliva, D.; Mohamed, W.M.; Hassaballah, M. A novel hybrid Harris hawks optimization and support vector machines for drug design and discovery. Comput. Chem. Eng. 2020, 133, 106656. [Google Scholar] [CrossRef]
  51. Bhuiyan, E.; Cerrai, D.; Biswas, N.K.; Zhang, H.; Wang, Y.; Chen, D.; Feng, D.; You, X.; Wu, W. Temperature Forecasting Correction Based on Operational GRAPES-3km Model Using Machine Learning Methods. Atmosphere 2022, 13, 362. [Google Scholar] [CrossRef]
  52. Valikhan Anaraki, M.; Farzin, S.; Mousavi, S.F.; Karami, H. Uncertainty Analysis of Climate Change Impacts on Flood Frequency by Using Hybrid Machine Learning Methods. Water Resour. Manag. 2021, 31, 199–223, 101007/s11269. [Google Scholar] [CrossRef]
Figure 1. Flow chart of the novel framework based on the SEML method for WS modeling.
Figure 1. Flow chart of the novel framework based on the SEML method for WS modeling.
Atmosphere 13 00758 g001
Figure 2. The schematic structure of ANN.
Figure 2. The schematic structure of ANN.
Atmosphere 13 00758 g002
Figure 3. The schematic structure of the LSSVM-GS.
Figure 3. The schematic structure of the LSSVM-GS.
Atmosphere 13 00758 g003
Figure 4. The schematic structure of the LSSVM-HHO.
Figure 4. The schematic structure of the LSSVM-HHO.
Atmosphere 13 00758 g004
Figure 5. Case study locations.
Figure 5. Case study locations.
Atmosphere 13 00758 g005
Figure 6. PDF plots of the observed values of WS at the different stations.
Figure 6. PDF plots of the observed values of WS at the different stations.
Atmosphere 13 00758 g006
Figure 7. Scatter plots of the observed and modeled values of WS using the meta-algorithm at different stations.
Figure 7. Scatter plots of the observed and modeled values of WS using the meta-algorithm at different stations.
Atmosphere 13 00758 g007
Figure 8. Time series of the observed and modeled values of WS using the meta-algorithm at different stations.
Figure 8. Time series of the observed and modeled values of WS using the meta-algorithm at different stations.
Atmosphere 13 00758 g008
Figure 9. Q–Q plots of observed and modeled values of WS using the meta-algorithm at different stations.
Figure 9. Q–Q plots of observed and modeled values of WS using the meta-algorithm at different stations.
Atmosphere 13 00758 g009
Table 1. Geographic information and statistical specifications of the data from the different stations.
Table 1. Geographic information and statistical specifications of the data from the different stations.
InputOutput
StationParameterTmin (°C)Tmean (°C)Tmax (°C)Hmin (%)Hmean (%)Hmax (%)SH (hr)WS (m/s)LongitudeLatitude
QuchanAverage7.8674.6455.1136.8319.6512.846.102.0658.5137.11
Std a3.3918.3419.7020.1210.469.638.071.37
SarakhsAverage8.2065.7848.8733.6023.9718.0210.262.2061.1636.53
Std a3.6319.2619.5518.9710.229.377.471.4
SabzevarAverage8.4458.9542.0226.8124.5217.4411.562.5457.6836.21
Std a2.6919.2916.5613.949.458.798.031.20
GolmakanAverage7.8765.9448.9134.4420.1415.076.572.9159.1536.48
Std a2.7617.7217.6917.259.318.796.711.65
MashhadAverage8.2971.5752.2833.9321.4414.768.502.1759.6136.3
Std a3.0519.7518.4517.139.508.517.360.97
NeyshaburAverage8.7566.0549.8032.2022.4015.596.751.3458.4536.45
Std a2.7517.9017.7316.369.168.656.570.79
Torbat HeydariehAverage9.0663.8145.6929.6321.2314.808.412.3859.2235.28
Std a3.1121.0017.7114.618.888.527.841.45
KashmarAverage9.0853.6639.4627.6924.1318.5911.501.5658.4735.33
Std a2.6921.1716.7013.159.098.727.590.94
GonabadAverage9.2154.5838.5427.0624.5619.2510.662.1558.5634.27
Std a2.5020.1016.8014.178.858.577.361.32
TabasAverage9.1145.3431.4319.9728.4421.7315.162.0556.9233.58
Std a2.6820.1715.5011.489.379.118.451.38
FerdowsAverage9.2952.9136.3724.2424.5718.6610.002.2058.1734.01
Std a2.5219.3216.1813.348.768.617.431.00
QaenAverage9.2255.8336.8924.5422.2416.626.922.6559.233.73
Std a2.5019.0015.2812.737.917.906.861.30
BirjandAverage9.1353.8535.9920.2124.1316.568.792.3859.1332.53
Std a2.4620.8316.1512.218.187.927.441.17
NehbandanAverage8.5748.1330.6319.8826.2120.7612.513.1760.0331.53
Std a2.7321.8615.5211.998.428.438.031.98
BoshruyehAverage8.4352.8236.5924.6927.1620.9011.431.3957.4533.87
Std a2.6118.3915.6513.109.339.107.811.00
Khor BirjandAverage9.1844.9128.4619.3226.0020.5712.494.9358.2632.56
Std a2.5118.2713.2110.768.538.778.102.26
a Standard deviation.
Table 2. Results of modeling WS with the base algorithms at the different stations.
Table 2. Results of modeling WS with the base algorithms at the different stations.
TrainingTesting
MAERMSERRMSERMAERMSERRMSER
Quchan
ANN0.951.230.790.610.961.230.800.61
MLR1.001.300.830.550.981.270.820.57
MNLR0.981.270.820.580.971.260.810.59
GRNN0.761.050.680.740.971.240.800.60
MARS0.961.240.800.610.951.230.800.61
M50.951.240.790.610.961.240.800.60
GBR0.871.130.730.690.951.220.790.61
RBFNN0.000.000.001.001.081.420.920.43
LSBoost0.931.200.770.630.951.230.790.61
LSSVM-GS0.821.070.680.740.951.220.790.61
LSSVM-HHO0.871.130.730.690.951.220.790.61
Torbat Heydarieh
ANN0.951.290.780.630.971.320.790.62
MLR1.001.350.810.581.011.380.820.57
MNLR0.981.340.800.591.001.360.810.59
GRNN0.751.070.650.771.001.360.810.59
MARS0.961.300.780.620.971.330.790.61
M50.961.300.790.620.991.350.810.59
GBR0.821.110.670.750.971.320.790.62
RBFNN0.000.000.001.001.341.801.070.34
LSBoost0.841.140.680.740.971.330.790.61
LSSVM-GS0.821.110.670.750.971.320.790.61
LSSVM-HHO0.821.110.670.750.971.320.790.62
Gonabad
ANN0.951.310.850.530.961.320.850.52
MLR1.001.360.880.481.001.360.880.48
MNLR0.991.350.870.491.001.350.870.48
GRNN0.751.080.700.730.981.350.870.49
MARS0.971.330.860.510.971.330.860.51
M50.961.320.850.520.971.330.860.51
GBR0.871.190.770.650.961.320.850.52
RBFNN0.000.000.001.001.151.490.970.27
LSBoost0.861.160.750.670.971.330.860.51
LSSVM-GS0.831.130.730.700.961.320.850.52
LSSVM-HHO0.871.190.770.650.961.320.850.52
Tabas
ANN0.821.110.870.490.841.140.880.48
MLR0.851.140.890.450.861.160.890.45
MNLR0.851.140.890.450.861.160.890.46
GRNN0.630.890.690.740.861.160.890.45
MARS0.831.120.880.480.851.150.880.48
M50.831.120.870.490.861.170.900.45
GBR0.400.710.450.890.400.710.470.62
RBFNN0.000.000.001.001.011.311.000.00
LSBoost0.771.040.810.590.841.150.880.48
LSSVM-GS0.700.950.740.700.841.140.880.48
LSSVM-HHO0.761.030.800.610.841.140.870.48
Table 3. Average results of the algorithms during the testing period at the sixteen stations.
Table 3. Average results of the algorithms during the testing period at the sixteen stations.
MAERMSERRMSER
ANN0.991.340.870.48
MLR1.021.370.890.44
MNLR1.011.360.880.46
GRNN1.011.360.890.45
MARS0.991.340.870.47
M51.001.360.880.46
GBR0.961.310.840.51
RBFNN1.291.661.100.12
LSBoost0.991.340.870.47
LSSVM-GS0.991.340.870.48
LSSVM-HHO0.991.330.870.49
Table 4. Results of modeling WS with LSBoost (meta-algorithm) at the different stations.
Table 4. Results of modeling WS with LSBoost (meta-algorithm) at the different stations.
TrainingTesting
MAERMSERRMSERMAERMSERRMSER
Quchan0.410.710.460.890.410.710.460.89
Sarakhs0.450.770.460.890.500.930.530.85
Sabzevar0.420.720.500.870.450.780.540.84
Golmakan0.561.110.550.840.571.200.600.80
Mashhad0.300.560.480.880.350.640.540.84
Neyshabur0.240.440.490.870.260.440.490.87
Torbat Heydarieh0.430.770.460.890.440.780.470.88
Kashmar0.300.550.490.870.320.600.530.85
Gonabad0.410.730.470.880.420.730.480.88
Tabas0.390.670.430.900.420.710.460.89
Ferdows0.330.600.510.860.340.610.510.86
Qaen0.410.720.470.880.420.760.500.87
Birjand0.300.550.400.920.370.690.520.86
Nehbandan0.561.040.460.890.581.060.480.88
Boshruyeh0.330.610.510.860.340.600.510.86
Khor Birjand0.691.230.480.880.711.290.490.87
Average0.410.740.480.880.430.780.510.86
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Morshed-Bozorgdel, A.; Kadkhodazadeh, M.; Valikhan Anaraki, M.; Farzin, S. A Novel Framework Based on the Stacking Ensemble Machine Learning (SEML) Method: Application in Wind Speed Modeling. Atmosphere 2022, 13, 758. https://0-doi-org.brum.beds.ac.uk/10.3390/atmos13050758

AMA Style

Morshed-Bozorgdel A, Kadkhodazadeh M, Valikhan Anaraki M, Farzin S. A Novel Framework Based on the Stacking Ensemble Machine Learning (SEML) Method: Application in Wind Speed Modeling. Atmosphere. 2022; 13(5):758. https://0-doi-org.brum.beds.ac.uk/10.3390/atmos13050758

Chicago/Turabian Style

Morshed-Bozorgdel, Amirreza, Mojtaba Kadkhodazadeh, Mahdi Valikhan Anaraki, and Saeed Farzin. 2022. "A Novel Framework Based on the Stacking Ensemble Machine Learning (SEML) Method: Application in Wind Speed Modeling" Atmosphere 13, no. 5: 758. https://0-doi-org.brum.beds.ac.uk/10.3390/atmos13050758

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop