Next Article in Journal
An NNwC MPPT-Based Energy Supply Solution for Sensor Nodes in Buildings and Its Feasibility Study
Next Article in Special Issue
Probabilistic Wind Power Forecasting Approach via Instance-Based Transfer Learning Embedded Gradient Boosting Decision Trees
Previous Article in Journal
Observation of Near-Inertial Oscillations Induced by Energy Transformation during Typhoons
Previous Article in Special Issue
An LM-BP Neural Network Approach to Estimate Monthly-Mean Daily Global Solar Radiation Using MODIS Atmospheric Products
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Using Smart Persistence and Random Forests to Predict Photovoltaic Energy Production

by
Javier Huertas Tato
1,* and
Miguel Centeno Brito
2,*
1
Department of Computer Science, Universidad Carlos III de Madrid, 28911 Madrid, Spain
2
Instituto Dom Luiz (IDL), Faculdade de Ciências, Universidade de Lisboa, 1749-016 Lisboa, Portugal
*
Authors to whom correspondence should be addressed.
Submission received: 29 November 2018 / Revised: 19 December 2018 / Accepted: 25 December 2018 / Published: 29 December 2018
(This article belongs to the Special Issue Solar and Wind Energy Forecasting)

Abstract

:
Solar energy forecasting is an active research problem and a key issue to increase the competitiveness of solar power plants in the energy market. However, using meteorological, production, or irradiance data from the past is not enough to produce accurate forecasts. This article aims to integrate a prediction algorithm (Smart Persistence), irradiance, and past production data, using a state-of-the-art machine learning technique (Random Forests). Three years of data from six solar PV modules at Faro (Portugal) are analyzed. A set of features that combines past data, predictions, averages, and variances is proposed for training and validation. The experimental results show that using Smart Persistence as a Machine Learning input greatly improves the accuracy of short-term forecasts, achieving an NRMSE of 0.25 on the best panels at short horizons and 0.33 on a 6 h horizon.

1. Introduction

Solar energy forecasting is a key issue in the development of renewable energy sources. Improving the reliability and accuracy of forecasts can increase the competitiveness of solar electricity in power markets. Several methods have been proposed in recent years aiming to improve the accuracy of forecasts [1] and the solar photovoltaic grid efficiency [2,3,4,5].
Machine learning is playing an important role in the development of many new technologies, and solar forecasting techniques are no exception. Forecasting solar resources requires addressing several challenges [6]. Many advances have been made applying machine learning techniques [7,8] to solar forecasting. The main advantage of machine learning algorithms is the automatic detection of hidden patterns in large amounts of data, allowing the identification of unknown synergies and integration of varied sources of information.
This article aims to combine simple radiation predictions with real radiation measurements to predict the solar photovoltaic energy production of a diverse variety of PV modules for short-term temporal horizons. Many proposals apply machine learning algorithms to predict both photovoltaic energy production and Solar Irradiance. Orjuela-Cañón et al. [9] compare several machine learning algorithms, both linear and nonlinear, such as Regression Trees, Artificial Neural Networks, and a Seasonal Autoregressive model. Gagne et al. [10] which compares Gradient Boosting and Random forests, both of which are combinations of simple models. Diagne et al. [11] review a wide variety of statistical models, including linear and nonlinear models such as Persistence, to Artificial Neural Networks. This analysis also includes the results of models based on cloud, satellite, and ground-based images. Numerical Weather Predictors (NWPs) are also compared. Finally, Pedro and Coimbra [12] evaluate more nonlinear models exclusively using radiation data, without weather forecasts.
Common techniques used for solar forecasting are Support Vector Regression (SVR) [13], Artificial Neural Networks (ANN) [14,15,16], hybrid approaches of ANN and SVR [17,18,19] or Vector Autoregression (VAR) [20].
NWP are commonly used [21,22,23] for daily forecasts. These studies show the powerful prediction capabilities of forecasts as machine learning inputs. Furthermore, NWPs are useful when forecasting other resources, such as wind. A set of wind power predictors can be used, combining them into secondary forecasts and finally reaching the final wind forecast [24]. Wang et al. [25] use ANN to predict wind velocity using an ensemble, ANN are used. This study predicts wind velocity Using an ensemble of neural networks (Back propagation, Wavelet, Radial basis function and Generalized regression neural networks).
A common technique used in solar forecasting is Smart Persistence. This assumes the solar energy will remain constant over time. At any given point in time t, Persistence will predict into the future h minutes ahead at the time point t + h as the PV energy measured in t. Therefore, Persistence follows this expression: P V ( t + h ) = P V ( t ) . The Persistence forecasting model is accurate on near horizons, but the accuracy decreases as the temporal horizon increases.
Smart Persistence is an improved version of Persistence which assumes the sky conditions will remain constant (instead of irradiance itself). The forecasting algorithm predicts the current solar radiation as the product of the current clear sky ratio ( K t ) and the clear sky radiation ( I c s ) in the predicted point. It is widely used as the baseline for validating more complex models and is fairly accurate in short-term horizons.
In the context of machine learning, Random Forests [26] is a classification and regression learning algorithm. A Random Forest is an ensemble learner. It trains a large number of simple decision trees and aggregates the results of every tree. Each one is trained using random subsets of the dataset to recognize different patterns and synergies in the data.
The Random Forest algorithm has been widely used in forecasting, for example Voyant et al. [7] reviews many articles in the context of machine learning applied to solar forecasting. Random Forests are rarely used for these problems, but they show very positive results whenever they are applied. The usage of this algorithm can be extended to PV production forecasting. When compared with other machine learning algorithms, Zamo [27] has argued that Random Forest (RF) is the most reliable option for forecasting PV production.
It is widely accepted that different machine learning algorithms perform differently based on the dataset used in the learning task. The RF algorithm has been chosen to forecast PV production. It has been used before in the literature and shows satisfying results on other datasets. It has been found to be the most suitable approach for the dataset tackled in this article.
This research explores the potential of introducing smart persistence predictions as inputs to machine learning forecasting techniques. The incremental methodology shown in this article demonstrates that using predictions along with past measures helps RFs find more accurate forecasts than the other methods proposed.
This article is structured as follows. First the dataset and methods used are detailed in Section 2. The features are described and then, the machine learning approach is explained. In Section 3 the experimentation is described, and the results are shown. The overall conclusions are presented in Section 4.

2. Materials and Methods

2.1. Data

The data used in this article has been provided by SunLab. The complete dataset has information on four solar stations from which Faro has been chosen. The Faro station has six different PV module models placed with optimal tilt. The data contains information on PV module temperature, meteorological variables, radiation, and PV production.
There are three years of data available with minute resolution. For short-term forecasting, minute-to-minute variations of radiation and production are very relevant. In this study, Global Horizontal Irradiance (GHI), Direct Normal Irradiance (DNI) and Photovoltaic production (PV) are used. The tested PV modules come from different manufacturers. Most of them have different PV production technology. There are six solar panel models tested labeled from A to F, panel A uses pc-Si technology, panel B uses CIS technology, panel C uses pc-Si technology, panel D uses pc-Si technology, panel E uses mc-Si technology, panel F uses CdS/CdTe thin film technology.

2.2. Method Overview

The PV production forecast problem addressed is a specific time series problem. It can be represented as the general expression for time series as presented in Equation (1).
P V ( t + h ) = f ( P V ( t 1 ) , P V ( t 2 ) , , P V ( t n ) )
In this equation, t is the time instant in minutes when the prediction is being made, h is the time horizon for which the production is being forecasted, and n is the number of time steps considered. The function f is learned and predicts PV production. In this context, the information for the t is not known, therefore any f will start with P V ( t 1 ) instead of P V ( t ) .
The general Equation (1) can be extended and further generalized to include further helpful information about weather, radiation, statistical measures, etc. This can be seen in Equation (2)
P V ( t + h ) = f ( P 1 ( t 1 ) , P 1 ( t 2 ) , , P 1 ( t n ) , P 2 ( t 1 ) , P 2 ( t 2 ) , , P 2 ( t n ) , , P m ( t 1 ) , P m ( t 2 ) , , P m ( t n ) )
The data point P i is a single predictor, and m is the number of predictors that will be used. This expanded expression allows the use of more information aside from the PV production data.
Using the proposed Equation (2), a set of statistical measures and instantaneous information will be used. PV production is forecasted on horizons from 15 to 360 min and take the following values: h = ( 15 , 30 , 45 , 60 , 120 , 180 , 240 , 300 , 360 ) . The time steps have been set as n = 3 . These features are described in further detail in Section 2.3.
To find an adequate predictor function f the RF algorithm is used as described in Section 2.4.

2.3. Feature Description

There are two categorizations for the features considered: time window (instantaneous or aggregate) or source (measured or predicted). Instantaneous features are those taken from a single minute of the dataset. The aggregate features are a statistical transformation (mean or standard deviation) of a set of individual minute data. The source categorization consists of measured data, taken exclusively by the instruments of the power plant; or predicted data, a linear transformation of a measure that produces a rough estimation of a future value.
Given these terms, in this study ten useful predictors have been considered: Photovoltaic production (PV), GHI, DNI, mean clear sky index ( K t ), standard deviation K t , PV Smart Persistence, mean PV Smart Persistence and Standard Deviation PV Smart Persistence. All of these are classified and sorted in Table 1.

2.3.1. PV Production

PV production is a measured and instantaneous feature, it is the energy production for a single minute of operation. It is represented as P V ( t ) . This measure is included in every model independently; it is never excluded because of the categories given.

2.3.2. GHI

GHI is a measured and instantaneous feature. It represents the amount of radiation received at a horizontal surface integrated over a single minute and its power measured as W/m 2 . It is represented as I g ( t ) .

2.3.3. DNI

DNI is a measured and instantaneous feature. It represents the total amount of radiation received at a surface perpendicular to the light source integrated over a single minute and its power is measured as W/m 2 . It is represented as I d ( t ) .

2.3.4. Mean K t

Mean clear sky ratio ( K t ) represents the mean K t ratio over a window of time, and it is a measured and aggregate feature. The clear sky radiation is calculated using the method proposed by European Solar Radiation Atlas (ESRA) [28]. It is commonly used as a baseline to compare more sophisticated models; however, in this research is it used to enhance the forecasting capabilities of RFs. K t is obtained by dividing I g ( t ) by the clear sky GHI, the theoretical maximum that I g ( t ) can be at any t point. This operation is shown in Equation (3), where I c s ( t ) represents the clear sky GHI.
K t ( t ) = I g ( t ) I c s ( t )
The Clear Sky ratio K t removes geometrical trends that the GHI measurement due to the position of the sun in the sky. The operation that gives the feature is presented in Equation (4).
K t ¯ ( t ) = 1 w i = 0 w K t ( t i )
where w represents the time window, in minutes. Here, it is always represented as w = 60 .

2.3.5. Standard Deviation K t

Standard deviation K t is the amount of variability of the K t ratio over a window of time, and it is a measured and aggregate feature. It is calculated as presented in Equation (5)
σ K t ( t ) = 1 w i = 0 w ( K t ( t i ) K t ¯ ( t i ) ) 2

2.3.6. PV Smart Persistence

PV smart persistence is the expected PV for a time point t + h if the sky was as cloudy then as it is in time t. It is an instantaneous and predicted feature. This feature is an estimator of the future and it is calculated as presented in Equation (6).
S P ( t , h ) = C · K t ( t ) · I c s ( t + h )
The constant C scales GHI to PV production value. It is estimated as the maximum PV power in a month, divided by the maximum GHI of the same month. This is an estimation and it is valid independently of the year it is used in. Hence, if these measures were taken for the year 2014, then it would be acceptable to reuse them again for the year 2015.
The horizon h is the minute for which the prediction is being made, and it is used in I c s ( t + h ) which is a value calculated without intervention of future knowledge.

2.3.7. Mean PV Smart Persistence

Mean PV Smart Persistence is the average value of Smart Persistence over an hour. It is an aggregate and predicted feature and it is calculated as presented in Equation (7)
S P ¯ ( t , h ) = 1 w i = 0 w S P ( t i , h )

2.3.8. Standard Deviation PV Smart Persistence

Standard deviation PV smart persistence is the amount of variability of the PV measures over a window of time w. It is an aggregate and predicted feature and it is calculated as presented in Equation (8).
σ S P ( t , h ) = 1 w i = 0 w ( S P ( t i , h ) S P ¯ ( t i , h ) ) 2

2.4. Machine Learning

The RF algorithm is an ensemble learner. It creates a set of decision trees that vote on a final result. Since forecasting is a regression problem, instead of decision trees, a set of regression trees are trained. The forecast will be the mean of the different regression tree results. The RF algorithm is used to build the f presented in Equation (2).
Every tree of the RF is trained using independent subsets of the dataset. Each subset is built selecting data randomly from the original input dataset. The number of trees must be specified in advance, as well as how much data is taken for each subset of the initial training set. The forecast of the RF is the average of the forecasts made by the regression trees.
To get the most reliable RF, the dataset must be split in training and testing datasets. The RF algorithm is trained and tested using different partitions of the data. The training set is used in the training and parameter tuning process of the RF learning. The test set is used to extract the final metrics. There exists a training and a test set for each PV module. The training set is represented by the first three weeks of each month and the test set is represented by the leftover data after the partition is made.
This algorithm requires the adjustment of its hyper-parameters. They are the number of trees, which has been set to 500, and the randomly selected learning examples. Tuning the hyper-parameters requires a validation set. The best hyper-parameter set is selected in three steps: (1) building several models with different hyper-parameter combinations using a subset of data from the training set, (2) evaluating the models using the unused subset of data from the training set (validation set) and (3) choosing the model with the lowest error metric on the validation set. To adjust the hyper-parameters of the algorithm, a brute force search is made for each input set from Table 1 with a random seed of 1.

3. Experimentation

To test the accuracy of the different features proposed an experimentation based on the feature classification presented in Table 1 is described in this section.
The hypothesis is that adding Smart Persistence information to the time series PV forecasting will improve the accuracy metrics in a significant way. To test this, the methods and features proposed in Section 2.3 and Section 2.4 respectively, are used following the experimental methodology of Section 3.1.

3.1. Method

The whole dataset contains information for six different PV modules, thus the data is divided in six different sets that share the same irradiance measures but are different in the PV measurements. To get the final errors measurements the metrics obtained at the experimentations are the average of all PV modules accuracy metrics.
A model is trained for each PV model to analyze the performance of the ML method independently of the panel model and technology. Then the results are averaged to obtain the overall performance.
There are two main metrics used in this study, the normalized Root Mean Squared Error (RMSE) and the Skill over Smart persistence. The NRMSE is a normalized measure of the RMSE. It is relative to the average measures from the real sample. The Skill is a relative measure comparing two RMSE errors. It represents how much percentage of Smart Persistence’s RMSE needs to be added (if positive) or subtracted (if negative) to get the RMSE of the evaluated method. Higher negative numbers indicate better skill. Both are represented in Equations (9) and (10) respectively. Aside from the two main metrics proposed, two additional metrics are evaluated, the RMSE in Equation (11) and the R squared in Equation (12)
NRMSE = 1 n j = 0 n ( p j d j ) 2 1 n j = 0 n d j
Skill = 1 n j = 0 n ( p j d j ) 2 1 n j = 0 n ( p j p e r s d j ) 2 1
RMSE = 1 n j = 0 n ( p j d j ) 2
RMSE = 1 j = 0 n ( p j d j ) 2 j = 0 n ( d j d ¯ ) 2
p j and d j represent pairs of predicted and measured values. p j p e r s represents the smart persistence prediction value. n is the number of prediction and observation pairs available.
Five different experiments are made: first dividing features by source, using only measured or predicted data, then dividing features by window, using only instantaneous or aggregate data. Finally, all features are included to build the final model. All experiments include the PV production feature.
Experiments are made with horizons up to 6 h. NWPs using Satellite and other forecasting methods are known to perform better beyond 6 h. The focus of this research is the nowcasting capabilities of integrating Smart Persistence.

3.2. Results

Results are shown in Figure 1, Figure 2 and Figure 3; and with Table 2 and Table 3. Figure 1 represents the NRMSE performance metric for every PV module available, Figure 2 shows the skill against performance. Here negative results correspond to the needed amount to subtract from persistence RMSE to get the RMSE of the experimented model and therefore more negative values correspond to higher forecasting accuracy. Figure 3 shows the average metrics of all PV modules. Table 2 compares two different forecasting methods, the RF approach, and a linear regression. Finally, Table 3 shows the comparison of all evaluated panel metrics, including R squared and RMSE. Both tables show results with the models constructed with every input.
It is observed that the NRMSE results are almost always lower on every PV module and horizon using the combination of all features. The exception is the PV module labeled as B, where this does not happen before the 3 h horizon. This happens both in NRMSE and Skill measurements. Excluding this exception, using all features is reliably the best option for all PV modules.
Exclusively using measurements is systematically the worst approach, being the least accurate model. It reaches very high NRMSE peaks and it is outperformed by Smart Persistence for middle horizons. The other models are frequently better than using only measured data. There is some variability on the magnitude of the metrics between PV modules. There is an ongoing trend on most panels: the best model uses every feature, followed by using instantaneous data, then aggregated data, then predicted data, and finally just measured data.
This means that using GHI, DNI and PV production just by themselves is worse than including Smart Persistence. Exclusively using predictions (both instantaneous and aggregate) is worse than using a combination of predictions and measure data. Using only aggregate or instantaneous data is worse than using a combination of both. All type of inputs presented to the RF contribute and complement each other, yielding the best results when all features are combined.
Skill metrics show similar trends to NRMSE; all PV modules behave similarly except B. However, the relative measures between PV modules vary significantly. The modules that perform best are both B and C with very high improvements. The other four PV modules evolve into worse Skill metrics, but then they start improving at the 3 h horizon. This happens because persistence is inaccurate at far horizons, but the machine learning model keeps the RMSE low enough to outperform smart persistence.
The magnitude of the Skill improvement remains very high in every model and PV module, except when using only measured data. For the best model, the improvement can go as high as 0.7 for near horizons and 0.6 for far horizons, while it never gets below 0.3 on any horizon.
The average of the metrics for each horizon shows that, even with the variance introduced by module B, using all features significantly improves results. The trends observed in individual are repeated for the averaged measurement. The main difference is that for the fifteen-minute horizon, the measured data model is just as good as the Aggregate and Instantaneous models.
Overall, the NRMSE metric appears to increase with the horizon, as is expected. It ends up doubling when using most methods. This pattern is shared by five out of the six solar panels tested. While NRMSE steadily rises through the horizons, most methods overcome Smart Persistence by a considerable margin, as shown by the Skill metric. The only time Smart Persistence is better than the ML approach is when the Measured input set is used. At the 6 h horizon, the Skill is better than at the 15 min horizon on most input sets. This is to be expected, as Smart persistence gets consistently worse as the prediction horizon increases.
A comparison between RF and linear regression (LM) is shown in Table 2. Results show that the machine learning algorithm outperforms the linear regression on all horizons and approaches. This shows that additional data by itself is not enough to get acceptable results and that machine learning significantly improves the results of forecasting. The skill metric for the linear model is high, meaning that the additional data improves predictions by a great margin independently of the model used.
Furthermore, in Table 3 all metric values are show for the best feature set. The magnitude of errors varies from panel to panel, which is very prominent on RMSE metrics. It is also shown that a high Skill metric does not convey a low error. The R squared metric for most panels (except B) is very similar when compared to the variance of the other metrics. Again, as with the Skill metric, a good R squared metric does not correlate with good prediction RMSE or nRMSE.

4. Conclusions

PV production forecasting is a relevant problem that is tackled in this article through the use of Smart Persistence as an input to a RF machine learning algorithm. Experimentation shows interesting new results. First of all, PV production forecasting depends on the PV modules analyzed. Training a RF with data from different modules will yield significantly different error metrics using the same methodology.
However, even with varied metrics, it has been shown that using only measured data is insufficient to accurately predict the future PV production for farther horizons. Estimations of future instants and trends (Smart Persistence) noticeably increases the accuracy of the machine learning models. Combining both predictions and measures has been found to be the best possible combination of features.
Further improvements could be achieved using a wider prediction window, including more accurate predictor data, or using more complex trend analysis. A wider prediction window with more instants of data might be able to improve predictions by using more relevant data. Including other powerful and more accurate predictors could guide the random forest algorithm closer to more accurate models. Finally, analyzing trends further, might allow the machine learning algorithm to estimate better predictions.
References

Author Contributions

Conceptualization, M.C.B.; methodology, J.H.T.; software, J.H.T.; validation, J.H.T. and M.C.B.; formal analysis, M.C.B.; investigation, J.H.T. and M.C.B.; resources, M.C.B.; data curation, J.H.T.; writing–original draft preparation, J.H.T.; writing–review and editing, J.H.T.; visualization, J.H.T.; supervision, M.C.B.; project administration, M.C.B.; funding acquisition, M.C.B.

Funding

This research was funded by FCT grant number UID/GEO/50019/2013. The researcher Javier Huertas Tato is funded by the Ministry of Science and Innovation (Spain) grant number ENE2014-56126-C2.

Acknowledgments

This research has been made possible by the data shared by the SUNLAB project from EDP and collaboration with the University of Lisbon. Javier Huertas is funded by project ENE2014-56126-C2 (Towards an integrated model for solar energy forecasting), funded by the Ministry of Science and Innovation (Spanish Government). The authors are grateful for the financial support provided by FCT thought project.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Inman, R.H.; Pedro, H.T.; Coimbra, C.F. Solar forecasting methods for renewable energy integration. Progress Energy Combust. Sci. 2013, 39, 535–576. [Google Scholar] [CrossRef]
  2. Melgar-Dominguez, O.D.; Pourakbari-Kasmaei, M.; Mantovani, J.R.S. Adaptive Robust Short-Term Planning of Electrical Distribution Systems Considering Siting and Sizing of Renewable Energy Based DG Units. IEEE Trans. Sustain. Energy 2019, 10, 158–169. [Google Scholar] [CrossRef]
  3. Bashir, A.A.; Pourakbari Kasmaei, M.; Safdarian, A.; Lehtonen, M. Matching of Local Load with On-Site PV Production in a Grid-Connected Residential Building. Energies 2018, 11, 2409. [Google Scholar] [CrossRef]
  4. Crossland, A.F.; Jones, D.; Wade, N.S.; Walker, S.L. Comparison of the Location and Rating of Energy Storage for Renewables Integration in Residential Low Voltage Networks with Overvoltage Constraints. Energies 2018, 11, 2041. [Google Scholar] [CrossRef]
  5. Home Ortiz, J.; Pourakbari Kasmaei, M.; Lopez, J.; Roberto Sanches Mantovani, J. A stochastic mixed-integer conic programming model for distribution system expansion planning considering wind generation. Energy Syst. 2018, 9. [Google Scholar] [CrossRef]
  6. Olowu, T.O.; Sundararajan, A.; Moghaddami, M.; Sarwat, A.I. Future Challenges and Mitigation Methods for High Photovoltaic Penetration: A Survey. Energies 2018, 11, 1782. [Google Scholar] [CrossRef]
  7. Voyant, C.; Notton, G.; Kalogirou, S.; Nivet, M.L.; Paoli, C.; Motte, F.; Fouilloy, A. Machine learning methods for solar radiation forecasting: A review. Renew. Energy 2017, 105, 569–582. [Google Scholar] [CrossRef]
  8. Terren-Serrano, G. Machine Learning Approach to Forecast Global Solar Radiation Time Series; The University of New Mexico: Albuquerque, New Mexico, 2016. [Google Scholar]
  9. Orjuela-Cañón, A.D.; Hernández, J.; Rivero, C.R. Very short term forecasting in global solar irradiance using linear and nonlinear models. In Proceedings of the 2017 IEEE Workshop on Power Electronics and Power Quality Applications (PEPQA), Bogota, Colombia, 31 May–2 June 2017; pp. 1–5. [Google Scholar] [CrossRef]
  10. Gagne, D.J.; McGovern, A.; Haupt, S.E.; Williams, J.K. Evaluation of statistical learning configurations for gridded solar irradiance forecasting. Sol. Energy 2017, 150, 383–393. [Google Scholar] [CrossRef]
  11. Diagne, M.; David, M.; Lauret, P.; Boland, J.; Schmutz, N. Review of solar irradiance forecasting methods and a proposition for small-scale insular grids. Renew. Sustain. Energy Rev. 2013, 27, 65–76. [Google Scholar] [CrossRef]
  12. Pedro, H.T.C.; Coimbra, C.F.M. Assessment of forecasting techniques for solar power production with no exogenous inputs. Sol. Energy 2012, 86, 2017–2028. [Google Scholar] [CrossRef]
  13. Das, U.K.; Tey, K.S.; Seyedmahmoudian, M.; Idna Idris, M.Y.; Mekhilef, S.; Horan, B.; Stojcevski, A. SVR-Based Model to Forecast PV Power Generation under Different Weather Conditions. Energies 2017, 10, 876. [Google Scholar] [CrossRef]
  14. Abdel-Nasser, M.; Mahmoud, K. Accurate photovoltaic power forecasting models using deep LSTM-RNN. Neural Comput. Appl. 2017. [Google Scholar] [CrossRef]
  15. Ceci, M.; Corizzo, R.; Fumarola, F.; Malerba, D.; Rashkovska, A. Predictive Modeling of PV Energy Production: How to Set Up the Learning Task for a Better Prediction? IEEE Trans. Ind. Inform. 2017, 13, 956–966. [Google Scholar] [CrossRef]
  16. Moncada, A.; Richardson, W.; Vega-Avila, R. Deep Learning to Forecast Solar Irradiance Using a Six-Month UTSA SkyImager Dataset. Energies 2018, 11, 1988. [Google Scholar] [CrossRef]
  17. Koca, A.; Oztop, H.F.; Varol, Y.; Koca, G.O. Estimation of solar radiation using artificial neural networks with different input parameters for Mediterranean region of Anatolia in Turkey. Expert Syst. Appl. 2011, 38, 8756–8762. [Google Scholar] [CrossRef]
  18. Ozgoren, M.; Bilgili, M.; Sahin, B. Estimation of global solar radiation using ANN over Turkey. Expert Syst. Appl. 2012, 39, 5043–5051. [Google Scholar] [CrossRef]
  19. Li, Z.; Rahman, S.M.; Vega, R.; Dong, B. A Hierarchical Approach Using Machine Learning Methods in Solar Photovoltaic Energy Production Forecasting. Energies 2016, 9, 55. [Google Scholar] [CrossRef]
  20. Bessa, R.J.; Trindade, A.; Miranda, V. Spatial-Temporal Solar Power Forecasting for Smart Grids. IEEE Trans. Ind. Inform. 2015, 11, 232–241. [Google Scholar] [CrossRef]
  21. Huang, L.X.; Isaac, G.A.; Sheng, G. Integrating NWP forecasts and observation data to improve nowcasting accuracy. Weather Forecast. 2012, 27, 938–953. [Google Scholar] [CrossRef]
  22. Lu, S.; Hwang, Y.; Khabibrakhmanov, I.; Marianno, F.J.; Shao, X.; Zhang, J.; Hodge, B.M.; Hamann, H.F. Machine learning based multi-physical-model blending for enhancing renewable energy forecast—Improvement via situation dependent error correction. In Proceedings of the 2015 European Control Conference (ECC), Linz, Austria, 15–17 July 2015; pp. 283–290. [Google Scholar] [CrossRef]
  23. Chen, Z.; Troccoli, A. Urban solar irradiance and power prediction from nearby stations. Meteorol. Z. 2017, 26, 277–290. [Google Scholar] [CrossRef]
  24. Sánchez, I. Adaptive combination of forecasts with application to wind energy. Int. J. Forecast. 2008, 24, 679–693. [Google Scholar] [CrossRef]
  25. Wang, J.; Heng, J.; Xiao, L.; Wang, C. Research and application of a combined model based on multi-objective optimization for multi-step ahead wind speed forecasting. Energy 2017, 125, 591–613. [Google Scholar] [CrossRef]
  26. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  27. Zamo, M.; Mestre, O.; Arbogast, P.; Pannekoucke, O. A benchmark of statistical regression methods for short-term forecasting of photovoltaic electricity production, part I: Deterministic forecast of hourly production. Sol. Energy 2014, 105, 792–803. [Google Scholar] [CrossRef]
  28. Rigollier, C.; Bauer, O.; Wald, L. On the clear sky model of the ESRA—European Solar Radiation Atlas—With respect to the Heliosat method. Sol. Energy 2000, 68, 33–48. [Google Scholar] [CrossRef]
Figure 1. NRMSE for all modules considering different inputs to the ML algorithm. Top row: NRMSE of panels A, B and C. Bottom row: NRMSE of panels D, E and F.
Figure 1. NRMSE for all modules considering different inputs to the ML algorithm. Top row: NRMSE of panels A, B and C. Bottom row: NRMSE of panels D, E and F.
Energies 12 00100 g001
Figure 2. Skill for all modules considering different inputs to the ML algorithm. Top row: NRMSE of panels A, B and C. Bottom row: NRMSE of panels D, E and F.
Figure 2. Skill for all modules considering different inputs to the ML algorithm. Top row: NRMSE of panels A, B and C. Bottom row: NRMSE of panels D, E and F.
Energies 12 00100 g002
Figure 3. Mean Performance of Experiments measured in Skill (Left) and NRMSE (Right).
Figure 3. Mean Performance of Experiments measured in Skill (Left) and NRMSE (Right).
Energies 12 00100 g003
Table 1. Feature classification.
Table 1. Feature classification.
InstantaneousAggregate
MeasuredPV productionMean k t
GHIStandard Deviation k t
DNI
PredictedPV smart persistenceMean SP
Standard Deviation SP
Table 2. Metric comparison of Random Forest against Linear Regression.
Table 2. Metric comparison of Random Forest against Linear Regression.
Modelh = 15h = 30h = 60h = 120h = 180h = 240h = 300h = 360Metric
lm31.2233.0435.7539.2741.6042.3341.8640.44RMSE
rf26.0428.6131.4635.2837.8138.3537.3335.29RMSE
lm0.3080.3200.3340.3480.3620.3730.3880.411nRMSE
rf0.2570.2770.2930.3120.3280.3380.3460.358nRMSE
lm−24.38−24.93−24.09−23.91−24.77−26.46−28.01−29.29Skill
rf−36.93−34.99−33.20−31.64−31.63−33.38−35.80−38.29Skill
lm0.8460.8030.7530.6950.6740.6760.6860.692R2
rf0.8710.8420.8040.7530.7300.7340.7510.766R2
Table 3. Panel metric numerical comparison.
Table 3. Panel metric numerical comparison.
Panelh = 15h = 30h = 60h = 120h = 180h = 240h = 300h = 360Metric
A26.0428.6131.4635.2837.8138.3537.3335.295RMSE
B16.4916.8417.8620.2720.2620.8722.4614.33RMSE
C20.7422.9425.6429.3832.2232.8632.0130.29RMSE
D25.6728.5131.7735.8138.9740.0038.8936.69RMSE
E26.2728.8731.7735.5138.0938.7737.7635.71RMSE
F9.7610.70511.82813.15213.51613.76313.29712.278RMSE
A0.2570.2770.2940.3130.3290.3390.3470.359nRMSE
B0.5580.560.5770.6240.6150.6480.7450.533nRMSE
C0.2830.3070.3290.3540.3780.390.3990.414nRMSE
D0.2580.2810.3020.3230.3450.3590.3670.379nRMSE
E0.2520.2720.2890.3080.3240.3350.3430.355nRMSE
F0.2550.2740.2930.3090.3130.3240.3310.337nRMSE
A−36.93−34.99−33.20−31.64−31.63−33.38−35.81−38.29Skill
B−62.75−62.37−60.54−55.54−54.69−51.56−44.89−62.08Skill
C−72.77−70.40−67.53−63.24−59.43−57.36−56.26−55.54Skill
D−47.81−44.79−41.57−38.24−35.71−35.19−36.47−38.08Skill
E−37.74−35.83−34.06−32.71−32.46−33.78−36.07−38.39Skill
F−42.53−39.84−36.87−34.34−34.45−35.01−37.24−39.91Skill
A0.8720.8420.8040.7540.730.7350.7510.766R2
B0.6270.6040.5730.5390.5370.5580.5790.606R2
C0.8920.8660.830.7770.7420.7430.7550.766R2
D0.8730.8410.7990.7470.7150.7150.7310.746R2
E0.870.840.8030.7550.7310.7340.750.764R2
F0.8750.8450.8110.7680.7660.7730.7880.807R2

Share and Cite

MDPI and ACS Style

Huertas Tato, J.; Centeno Brito, M. Using Smart Persistence and Random Forests to Predict Photovoltaic Energy Production. Energies 2019, 12, 100. https://0-doi-org.brum.beds.ac.uk/10.3390/en12010100

AMA Style

Huertas Tato J, Centeno Brito M. Using Smart Persistence and Random Forests to Predict Photovoltaic Energy Production. Energies. 2019; 12(1):100. https://0-doi-org.brum.beds.ac.uk/10.3390/en12010100

Chicago/Turabian Style

Huertas Tato, Javier, and Miguel Centeno Brito. 2019. "Using Smart Persistence and Random Forests to Predict Photovoltaic Energy Production" Energies 12, no. 1: 100. https://0-doi-org.brum.beds.ac.uk/10.3390/en12010100

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop