Next Article in Journal
Production of Oxygenated Fuel Additives from Residual Glycerine Using Biocatalysts Obtained from Heavy-Metal-Contaminated Jatropha curcas L. Roots
Previous Article in Journal
Application of the Impedance Spectroscopy as a New Tool for Studying Biodiesel Fuel Aging Processes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Electric Energy Consumption Prediction by Deep Learning with State Explainable Autoencoder

Department of Computer Science, Yonsei University, Seoul 03722, Korea
*
Author to whom correspondence should be addressed.
Submission received: 3 February 2019 / Revised: 19 February 2019 / Accepted: 19 February 2019 / Published: 22 February 2019
(This article belongs to the Section A1: Smart Grids and Microgrids)

Abstract

:
As energy demand grows globally, the energy management system (EMS) is becoming increasingly important. Energy prediction is an essential component in the first step to create a management plan in EMS. Conventional energy prediction models focus on prediction performance, but in order to build an efficient system, it is necessary to predict energy demand according to various conditions. In this paper, we propose a method to predict energy demand in various situations using a deep learning model based on an autoencoder. This model consists of a projector that defines an appropriate state for a given situation and a predictor that forecasts energy demand from the defined state. The proposed model produces consumption predictions for 15, 30, 45, and 60 min with 60-min demand to date. In the experiments with household electric power consumption data for five years, this model not only has a better performance with a mean squared error of 0.384 than the conventional models, but also improves the capacity to explain the results of prediction by visualizing the state with t-SNE algorithm. Despite unsupervised representation learning, we confirm that the proposed model defines the state well and predicts the energy demand accordingly.

1. Introduction

As industrialization has progressed globally and the industry has developed, the demand for energy has become so high that energy has become an important topic in national policy [1]. In addition, energy use is rapidly increasing due to economic growth and human development [2]. The causes of these phenomena can be attributed to uncontrolled energy use such as overconsumption, poor infrastructure, and wastage of energy [3]. Among the demanders of various energy sources, Streimikiene estimates that residential energy consumption will account for a large proportion by 2030 [4]. According to Zuo, 39% of the United States’ total energy use is referred to as building energy consumption [5]. An energy management system (EMS) like a smart grid has been proposed to control the demand for soaring energy.
One work cycle of the EMS is the Plan-Do-Check-Act (PDCA) cycle as depicted in Figure 1 [6]. Formulating an energy plan is the first thing to do. This is the decision of the initial energy baseline, the energy performance indicators, the strategic and operative energy objectives, and the action plans. In the “do” phase, planning and action take place. The plans conducted in the previous phase have to be checked to ensure that they are effective. In the last phase, the results are reviewed, and a new strategy is established. Among the four stages, the “plan” phase is very important because it is the stage of establishing an energy use strategy and it includes an energy demand forecasting step. Therefore, it is necessary to study the energy prediction model to construct an efficient EMS.
Many researchers have conducted studies with various methods to predict energy demand. In the past, machine learning techniques such as the support vector machine (SVM) and linear regression (LR) have been widely used. However, as shown in Figure 2a, energy demand values over time are complex and noisy, which limits performance. As depicted in Figure 2b, the Fourier transform to analyze patterns of energy demand reveals that it has complex features. For quantitative analysis, t-test and ANOVA were performed on the dataset used in this paper as shown in Table 1. In the statistical analysis using the t-test, two groups (e.g., two different months in monthly demand) are chosen randomly and computed p-values, and we compute the average of all possible sampling. In the case of using ANOVA, p-value is computed from all the groups in each month, date, and hour.
As the result of our analysis that the characteristics of energy demand are complex, we have conducted the research with deep learning to extract difficult characteristics and work out tasks. Studies conducted using deep learning have contributed a lot to improving prediction performance, but they have not led to more utility from an energy management system perspective. If the energy demand forecasting model in the EMS copes with various situations and predicts the demand, it will be able to build a more efficient system. Therefore, in this paper, we propose a model that predicts future energy demand with that until now, considers various situations, and predicts energy demand according to different situations. This model consists of a projector that defines the state based on the energy demand to date and a predictor that predicts future energy demand from the defined state. We can adjust the automatically learned state in the middle to predict the energy demand with the consideration of various situations. The summary of the main contribution is as follows.
  • We propose a novel predictive model that can be explained by not only predicting future demand for electric power but also defining current demand pattern as state.
  • Our model predicts a very complex power demand value with stable and high performance compared with previous studies.
  • We analyze the state defined in latent space by the proposed model and investigate a model that predicts the power demand by assuming various explanations.
The rest of the paper is as follows. Section 2 introduces the previous studies for forecasting energy demand and addresses the limitations. To overcome these shortcomings, we propose our model in Section 3. Section 4 shows the results of the energy demand forecasting with the proposed model and also shows the result of forecasting the demand by considering various situations. In the final section, conclusions and discussion are presented.

2. Related Works

Several studies have been conducted to predict energy demand mentioned in Section 1. Table 2 summarizes the previous studies. In the past, statistical techniques were used mainly to predict energy demand. Munz et al. predicted a time series of irregular patterns using k-means clustering [7]. Kandananond used different forecasting methods—autoregressive integrated moving average (ARIMA), artificial neural network (ANN), and multiple linear regression (MLR) —to predict energy consumption [8]. Cauwer et al. proposed a method to predict energy consumption using a statistical model and its underlying physical principles [9].
However, due to the irregular patterns of energy demand, statistical techniques have limited performance and many models of prediction using machine learning methods have been investigated. Dong et al. predicted the demand of building energy using SVM with consumption and weather information [10]. Gonzalez and Zamarreno forecasted the next temperature from the temperature to date using a feedforward neural network (NN) and predicted the requirement with the difference of them [11]. Ekici and Aksoy predicted the building energy needs with properties of buildings without weather conditions [12]. Li et al. estimated the annual energy demand using SVM with the building’s transfer coefficient [13]. However, these studies only constructed models to predict correct value corresponding to the input so as to lack the basis for influence of the input features. To solve this problem, Xuemei et al. set the state for forecasting energy consumption through fuzzy c-means clustering and predicted demand with fuzzy SVM [14]. Ma forecasted energy consumption with specific population activities or unexpected events, as well as weather condition as inputs of the MLR model [15]. Although the above studies set the state and forecasted future consumption based on it, they lacked the mechanism to identify the state accurately.
As mentioned in Section 1, the energy consumption data contain large noise. Deep learning, which is a rising method to solve complex tasks of late, is efficient for predicting energy demand because it solves tasks by modeling complex characteristics of data well [16]. Ahmad et al. forecasted energy demand by constructing a deep NN and inputting the information of weather and building usage rate [17]. Lee et al. estimated environmental consumption by using a temporal model like recurrent neural network (RNN) with energy consumption data and temporal features [18]. Li et al. proposed a method to predict energy demand with autoencoder, one of the methods to represent data [19]. However, Li et al.’s model included only fully-connected layers, so that temporal features were ignored, and it is hard to control the conditions because latent space where the features of data are represented is not defined in that model.
Although some of the above studies provided novel research directions, other features such as information of weather and building are used in addition to the energy demand value, which is costly to construct the model for energy consumption prediction. Besides, they lack the explanation capability on the predicted value, because there was no study on the state to analyze the results of prediction. However, studies that analyze and explain the predictive results are essential for practical use of the predictive model. In this paper, we propose a model that visualizes a state by defining the state based on the current usage pattern and date information, so as to be able to explain the results of prediction. Our model, like any other prediction models, takes the energy demand up to now as input and predicts consumption in the future. However, in order to overcome the limitations of the end-to-end system, which cannot analyze the internal prediction process, we add a step to define the state of the demand pattern in the middle.

3. Proposed Method

3.1. Overview

As mentioned in Section 1 and Section 2, nonlinear approaches, including those based on fuzzy and neural net, have demonstrated successful performance in many applications [20,21,22,23]. In this paper, we have contributed to the field of application by solving the power demand forecasting problem using the deep neural network-based method. Compared to the previous work, the overall architecture of our model consists of a projector f and a predicter g , similar to an auto-encoder consisting of an encoder and a decoder as shown in Figure 3 [24]. There are many ways to deal with time series data, but f and g are based on long short-term memory (LSTM), one of the RNN’s, to handle time series data [25,26,27,28]. Predictor uses the output value of each time-step as the input of the next [29]. The projector defines the state by compressing the energy demand and transferring it to the latent space representing the demand information. Predictor predicts future energy demand based on the defined state. It can be done by end-to-end learning of the projector and the predictor, and the process of defining state is trained automatically. In the variational autoencoder (VAE), the state defined on the latent space contains the feature of the produced data, and also contains the information of the expected energy consumption, as well as features of the input values [30,31]. Ma and Lee predicted the energy consumption by adding more information of the surrounding environment while learning [15,18]. However, unlike them, after learning to predict the consumption with only demand to date, our model can predict future consumption by adjusting the state on the latent space with the condition of the surrounding environment.

3.2. Consumption Representation

Previous studies have constructed a model that sends energy consumption X to a predicted value Y as in Equation (1), while we define a state by adding a potential space S between X and Y , as shown in Equation (2).
f : X Y
f : ( S , X ) S
We continuously update the state during the time interval t. The state s i e for i during the time interval t of input is defined as follows.
s i e = f ( x i e , s i 1 e ) ,
where e means “embedded state”, x i e is the ith input value of projector and s 0 e = 0 . f ( , ) is a LSTM for the projector including one memory cell c t and three gates (input i t , forget f t , and output o t ). We calculate each value as follows.
s t e = o t t a n h c t ,
o t = σ ( U o x t e + W o s t 1 e ) ,
c t = f t c t 1 + I t c ¯ t ,
I t = σ ( U I x t e + W I s t 1 e ) ,
f t = σ ( U f x t e + W f s t 1   e ) ,
c ¯ t = t a n h ( U c x t e + W c s t 1 e ) ,
where U and W are weights of the layer, σ is activation function, and c ¯   and c are intermediary memory cell and memory cell, respectively. We set s t e as the last state of the projector (i.e., s t e = f ( x t e ,   s t 1 e ) ). The process to define the state is learned automatically as the predictor interacts with the projector to predict the demand. The details are introduced in the next section.
The state of the projector is located on the latent space where patterns and features of input energy consumption are shown. Therefore, by controlling the state transferred to the latent space, it is possible to predict the future consumption as well as to analyze the current consumption situation.

3.3. Demand Prediction

This section presents how to use the state set by the projector for forecasting future demand. First, as shown in Equations (10) and (11), the predictor predicts a single consumption value immediately after inputting the state. Recursively, predictor forecasts the next single demand value with the predicted value.
g : ( Y , S ) Y ,
s i d = g ( y i 1 , s i 1 d ) ,
where d means “demand”, y i is the ith output and the (i + 1)th input of predictor at the same time, and y 0 = 0 . g ( , ) is a RNN for the predictor. It gets the output of itself from the previous time-step as input and is computed as follows.
s i d = σ ( U s y i 1 + W s s i 1 d + b s )
y i = U y s i d + b y
Since the predictor forecasts the demand based on the first input state, if the state is arbitrarily controlled by the user, the model can predict the demand according to various conditions, resulting in an effective energy management system. For example, adding different conditions, such as information of weather or economy, can make energy demand prediction model more smoothly.
To train the proposed model, we use L2 loss function as shown in Equation (14) by sampling the data with a time interval of t and T in X and Y , respectively. The projector interacts with the predictor and learns to automatically define the state.
= E ( x 1 : t , y 1 : T ) ( X , Y ) [ ( y 1 : T g ( f ( x 1 : t ) ) ) 2 ]
The algorithms for learning the proposed model and forecasting the future demand are as follows. In Algorithm 1, two functions f and g are trained with energy demand record for t minutes,   x 1 : t , and energy demand to be predicted for T minutes, y 1 : T . In Algorithm 2, at first, we get a state with energy consumption x 1 : t and the predictor f . Then, based on the computed state s t e , we predict total T minutes by 1 min with the predictor g .
Algorithm 1. Learning algorithm for the proposed method
  • Input: Dataset ( X , Y )
  • Output: Trained model f ,   g
  • Initialize projector f and predictor g ;
  • repeat
  •  Sample x 1 : t ,   y 1 : T from X ,   Y , respectively;
  • θ f θ f θ f ( ) ;
  • θ g θ g θ g ( ) ;
  • until θ f , θ g converge
  • return f , g
where θ f , θ g are parameters of projector   f and predictor g , respectively.
Algorithm 2. An algorithm to predict electric energy demand
  • Input: x 1 : t
  • Output: y 1 : T
  • for i = 1 , t do
  • s i e f ( x i , s i 1 e ) ;
  • end for
  • s 1 d f d ( y 0 , s t e ) ;
  • y 1 U y s 1 d + b y ;
  • for i = 2 , , T do
  • s i d g ( y i 1 , s i 1 d ) ;
  • y i U y s i d + b y ;
  • end for
  • return y 1 : T

3.4. State Transition

This section describes how the user can arbitrarily adjust the defined state to assume various situations and predict energy demand. When energy demand x A based on condition A (e.g., warm weather) comes in as input, and when predicting energy demand based on condition B (e.g., cold weather), first, the average of state s A ¯ of the demanded quantities of condition A and the average s B ¯ of states of the demanded quantities of condition B are calculated. If the state s A , which is defined by inputting x A to projector, is subtracted from s A ¯ and s B ¯ is added to the state calculated through the energy demand x A , it can be seen that the condition A is omitted, and the condition B is added. We perform the state transition with this process and forecast the energy demand according to various conditions.

4. Experiments

4.1. Dataset amd Experimental Settings

To verify the proposed model, we use a dataset on household electric power consumption [32]. There are about two million minutes of electric energy demand data from 2006 to 2010, and they are divided into training and test data as a 9:1 ratio. It consists of eight attributes including date, global active power (GAP), global reactive power (GRP), global intensity (GI), voltage, sub metering 1, 2, and 3 (S1, 2, and 3), and the model predicts the GAP. S1 corresponds to the kitchen, containing mainly a microwave, an oven, and a dishwasher. S2 corresponds to the laundry room, containing a refrigerator, a tumble-drier, a light, and a washing-machine. S3 corresponds to an air-conditioner and an electric water-heater. The statistical summary of each feature is described in Table 3.
To train the time series model, we use a backpropagation through time (BPTT) algorithm, and the Adam optimizer with default hyper parameters in the keras library of python [33,34]. All weights are initialized with Glorot initialization [35]. The operating system of the computer used in our experiments was Ubuntu 16.04.2 LTS and the central processing unit of the computer was an Intel Xeon E5-2630V3. The random-access memory of the computer was Samsung DDR4 16 GB × 4, and the graphic processing unit of the computer was GTX Titan X D5 12 GB. The number of hidden units in the deep learning approach, including our model (i.e., the size of the state s ) was set at 64.

4.2. Demand Prediction

To verify the performance of the proposed model, we show the energy demand forecasting result using our model and compared with other conventional methods. Figure 4 is the result showing real and predicted energy demand values at the same time. The model predicts energy demand for 15, 30, 45, and 60 min with actual energy demand for 60 min. Although the model could not predict the energy demand perfectly, we confirm that the energy demand pattern predicted well. We show the convergence of the learning algorithm experimentally by showing the change of loss value as learning progresses in Figure 5.
Our model is compared with conventional machine learning methods such as linear regression (LR), decision tree (DT), random forest (RF) and multilayer perceptron (MLP), and with deep learning methods such as LSTM, stacked LSTM, and the autoencoder model proposed by Li. Stacked LSTM is a model including two LSTM layers similar to our model but does not set the state. A model proposed by Li has one hundred hidden units and four hidden layers. The MSE measure of the experimental results for each model is shown in Figure 6 as box plot. The results of the comparison with other models show that the proposed model outperforms other models. We can confirm that the conventional machine learning methods (LR, DT, RF, and MLP) show a large variation in prediction performance, but the deep learning methods (LSTM, Stacked LSTM, the Li’s model, and ours) are trained in stable. Some of the deep learning methods are worse than machine learning methods, but our model yields the best performance.
To examine the performance of the prediction model, we use three evaluation metrics—the mean squared error (MSE), the mean absolute error (MAE), and the mean relative error (MRE), which can be calculated respectively as follows.
M S E = 1 N Σ i = 1 N ( y ^ m y m ) 2
M A E = 1 N Σ i = 1 N | y ^ m y m |
M R E = 1 N Σ i = 1 N | y ^ m y m | y m
We conducted the experiments with 10-fold cross validation and the average values for each metric are shown in Table 4.

4.3. State Transition

We empirically verify whether our model can automatically learn a capacity to define state as described in Section 3. We extract the output of the projector to get states and visualize them as shown in Figure 7. We use the t-SNE algorithm to visualize the state [36]. We confirm that the consumption data are not separated clearly by month, but they are clustered by month on the latent space even with the unsupervised representation learning. Approximately the distribution of data can be divided into right (January, February, May, October, and November), left (August and September), top (December), center (March), center-right (June), center-left (July), and center-top (April). We mark plotted points of each month with annotations to figure and illustrate it monthly to show the state for each month, resulting in twelve plots. It can be seen that the defined state is well clustered on a monthly basis, achieving low intra-class variability.
Our model is also empirically confirmed that not only defines the state well but also has the capacity to adjust the prediction by controlling the state on latent space. This method is effective for EMS, for example, because the electricity demand prediction can be made flexible according to the climate or economic situation. The experiment to control the condition of the month is conducted, and other conditions are left for future study. An example of a state transition that controls a state on the latent space is shown in Figure 8. If we want to forecast the energy demand in October with only consumption in April, we just project the demand in April into the latent space to extract the state. After extracting the state for the electric energy consumption at one point in April, we add the average value of the states for October v ( x O C T ¯ ) and subtract the average value of the states for the April v ( x A P R ¯ ) and put it into the predictor to get the predicted values.
Table 5 shows the average value of the predicted electric power consumption for one hour in minutes to determine whether the demand pattern for April is changed to the consumption pattern for October after the state transition. Each column shows the month of the input electric energy demand to date, and each row shows an output month of state transition to predict the desired pattern for the specified month. The ground truth (GT) is the average electric energy consumption for each month. It can adjust the state on the latent space because the predicted consumption after conditioning is similar to GT.

5. Conclusions

We have addressed the importance of energy demand prediction and proposed a model to solve them. It attempts to predict electric energy consumption through defining the state unlike the conventional machine learning or deep learning models. Divided into two parts (projector and predictor), each part interacts with the other and learns to automatically set a state without any supervision. We achieve the best forecasting performance compared to others, analyze the state, and peek the basis for the predicted consumption value. In addition, the state transition method shows that our model can be more efficient because we can control the predicted electric energy consumption values according to various situations by adjusting conditions. For example, if we add several conditions to the state such as information of weather or economy, we can predict electricity demand accordingly.
In this paper, we have conducted experiments with several conditions of the state only for the month. We will experiment with various conditions such as weather, economy, or any other events in the future works. In this paper, only the energy consumption of one individual household is predicted, so the demand of several buildings will be collected, and we will add the information about building into the state and have a plan to propose a model which can predict energy consumption of various buildings. Finally, we will construct an efficient energy management system including the proposed prediction model.

Author Contributions

All authors have participated in preparing the research from the beginning to the end, such as establishing research design, method and analysis. All authors discussed and finalized the analysis results to prepare manuscript according to the progress of research.

Acknowledgments

This research was supported by Korea Electric Power Corporation (Grant number: R18XA05).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Oluklulu, Ç. A Research on the Photovoltaic Modules That Are Being Used Actively in Utilizing Solar Energy, Sizing of the Modules and Architectural Using Means of the Modules. Master’s Thesis, Gazi University, Ankara, Turkey, 2001. [Google Scholar]
  2. Ugursal, V.I. Energy Consumption, Associated Questions and some Answers. Appl. Energy 2014, 130, 783–792. [Google Scholar] [CrossRef]
  3. Rinkesh. What is the Energy Crisis. Available online: https://www.conserve-energy-future.com/causes-and-solutions-to-the-global-energy-crisis.php#abh_posts (accessed on 25 January 2019).
  4. Streimikiene, D. Residential Energy Consumption Trends, Main Drivers and Policies in Lithuania. Renew. Sustain. Energy Rev. 2014, 35, 285–293. [Google Scholar] [CrossRef]
  5. Zuo, J.; Zhao, Z.Y. Green Building Research-Current Status and Future Agenda: A review. Renew. Sustain. Energy Rev. 2014, 30, 271–281. [Google Scholar] [CrossRef]
  6. Prashar, A. Adopting PDCA (Plan-Do-Check-Act) Cycle for Energy Optimization in Energy-Intensive SMEs. J. Clean. Prod. 2017, 145, 277–293. [Google Scholar] [CrossRef]
  7. Munz, G.; Li, S.; Carle, G. Traffic Anomaly Detection Using k-means Clustering. In Proceedings of the GI/ITG Workshop MMBnet, Hamburg, Germany, 13–14 September 2007; pp. 13–14. [Google Scholar]
  8. Kandananond, K. Forecasting Electricity Demand in Thailand with an Artificial Neural Network Approach. Energies 2011, 4, 1246–1257. [Google Scholar] [CrossRef]
  9. De Cauwer, C.; Van Mierlo, J.; Coosemans, T. Energy Consumption Prediction for Electric Vehicles based on Real-world Data. Energies 2015, 8, 8573–8593. [Google Scholar] [CrossRef]
  10. Dong, B.; Cao, C.; Lee, S.E. Applying Support Vector Machines to Predict Building Energy Consumption in Tropical Region. Energy Build. 2005, 37, 545–553. [Google Scholar] [CrossRef]
  11. Gonzalez, P.A.; Zamarreno, J.M. Prediction of Hourly Energy Consumption in Buildings Based on a Feedback Artificial Neural Network. Energy Build. 2005, 37, 595–601. [Google Scholar] [CrossRef]
  12. Ekici, B.B.; Aksoy, U.T. Prediction of Building Energy Consumption by Using Artificial Neural Networks. Adv. Eng. Softw. 2009, 40, 356–362. [Google Scholar] [CrossRef]
  13. Li, Q.; Ren, P.; Meng, Q. Prediction Model of Annual Energy Consumption of Residential Buildings. In Proceedings of the 2010 International Conference on Advances in Energy Engineering, Beijing, China, 19–20 June 2010; pp. 223–226. [Google Scholar]
  14. Xuemei, L.; Yuyan, D.; Lixing, D.; Liangzhong, J. Building Cooling Load Forecasting Using Fuzzy Support Vector Machine and Fuzzy C-mean Clustering. In Proceedings of the 2010 International Conference on Computer and Communication Technologies in Agriculture Engineering, Chengdu, China, 12–13 June 2010; pp. 438–441. [Google Scholar]
  15. Ma, Y.; Yu, J.Q.; Yang, C.Y.; Wang, L. Study on Power Energy Consumption Model for Large-scale Public Building. In Proceedings of the 2010 2nd International Workshop on. IEEE Intelligent Systems and Applications, Wuhan, China, 22–23 May 2010; pp. 1–4. [Google Scholar]
  16. LeCun, Y.; Bengio, Y.; Hinton, G. Deep Learning. Nature 2015, 521, 436. [Google Scholar] [CrossRef] [PubMed]
  17. Ahmad, M.W.; Mourshed, M.; Rezgui, Y. Trees vs Neurons: Comparison between Random Forest and ANN for High-resolution Prediction of Building Energy Consumption. Energy Build. 2017, 147, 77–89. [Google Scholar] [CrossRef]
  18. Lee, D.; Kang, S.; Shin, J. Using Deep Learning Techniques to Forecast Environmental Consumption Level. Sustainability 2017, 9, 1894. [Google Scholar] [CrossRef]
  19. Li, C.; Ding, Z.; Zhao, D.; Yi, J.; Zhang, G. Building Energy Consumption Prediction: An Extreme Deep Learning Approach. Energies 2017, 10, 1525. [Google Scholar] [CrossRef]
  20. Herrera-Viedma, E.; Lopez-Herrera, A.G. A Review on Information Accessing Systems based on Fuzzy Linguistic Modelling. Int. J. Comput. Intell. Syst. 2010, 3, 420–437. [Google Scholar] [CrossRef]
  21. Haidegger, T.; Kovacs, L.; Precup, R.E.; Benyo, B.; Benyo, Z.; Preitl, S. Simulation and Control for Telerobots in Space Medicine. Acta Astronaut. 2012, 81, 390–402. [Google Scholar] [CrossRef]
  22. Saadat, J.; Moallem, P.; Koofigar, H. Training Echo State Neural Network Using Harmony Search Algorithm. Int. J. Artif. Intell. 2017, 15, 163–179. [Google Scholar]
  23. Ruiz-Rangel, J.; Hernandez, C.J.A.; Gonzalez, L.M.; Molinares, D.J. ERNEAD: Training of Artificial Neural Networks based on a Genetic Algorithm and Finite Automata Theory. Int. J. Artif. Intell. 2018, 16, 214–253. [Google Scholar]
  24. Kim, T.Y.; Cho, S.B. Predicting the Household Power Consumption Using CNN-LSTM Hybrid Networks. In Proceedings of the Int. Conf. on Intelligent Data Engineering and Automated Learning, Madrid, Spain, 21–23 November 2018; pp. 481–490. [Google Scholar]
  25. Ronao, C.A.; Cho, S.B. Human Activity Recognition with Smartphone Sensors Using Deep Learning Neural Networks. Expert Syst. Appl. 2016, 59, 235–244. [Google Scholar] [CrossRef]
  26. Kim, K.H.; Cho, S.B. Modular Bayesian Networks with Low-Power Wearable Sensors for Recognizing Eating Activities. Sensors 2017, 17, 2877. [Google Scholar]
  27. Kim, T.Y.; Cho, S.B. Web Traffic Anomaly Detection Using C-LSTM Neural Networks. Expert Syst. Appl. 2018, 106, 66–76. [Google Scholar] [CrossRef]
  28. Hochreiter, S.; Schmidhuber, J. Long Short-term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  29. Cho, K.; Van Merrienboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning Phrase Representations Using RNN Encoder-decoder for Statistical Machine Translation. arXiv, 2014; arXiv:1406.1078. [Google Scholar]
  30. Kingma, D.P.; Welling, M. Auto-encoding Variational Bayes. arXiv, 2013; arXiv:1312.6114. [Google Scholar]
  31. Larsen, A.B.L.; Sonderby, S.K.; Larochelle, H.; Winther, O. Autoencoding beyond Pixels Using a Learned Similarity Metric. arXiv, 2015; arXiv:1512.09300. [Google Scholar]
  32. Dua, D.; Karra, T.E. UCI Machine Learning Repository; University of California, School of Information and Computer Science: Irvine, CA, USA, 2007; Available online: http://archive.ics.uci.edu/ml (accessed on 21 February 2019).
  33. Hochreiter, S.; Bengio, Y.; Frasconi, P.; Schmidhuber, J. Gradient Flow in Recurrent Nets: The difficulty of Learning Long-term Dependencies. In A Field Guide to Dynamical Recurrent Neural Networks; IEEE Press: Piscataway, NJ, USA, 2001. [Google Scholar]
  34. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv, 2014; arXiv:1412.6980. [Google Scholar]
  35. Glorot, X.; Bengio, Y. Understanding the Difficulty of Training Deep Feedforward Neural Networks. In Proceedings of the 13th International Conference on Artificial Intelligence and Statistics, Sardinia, Italy, 13–15 May 2010; pp. 249–256. [Google Scholar]
  36. Maaten, L.V.D.; Hinton, G. Visualizing Data Using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
Figure 1. Plan, do, check, and act (PDCA) cycle for EMS. We focus on “plan” phase in this paper.
Figure 1. Plan, do, check, and act (PDCA) cycle for EMS. We focus on “plan” phase in this paper.
Energies 12 00739 g001
Figure 2. (a) The electric energy demand for each date, and (b) the result of Fourier transform.
Figure 2. (a) The electric energy demand for each date, and (b) the result of Fourier transform.
Energies 12 00739 g002
Figure 3. The overall scheme of the proposed method.
Figure 3. The overall scheme of the proposed method.
Energies 12 00739 g003
Figure 4. The predicted electric energy consumption and the actual demand. We show the prediction results for (a) 15, (b) 30, (c) 45, and (d) 60 min.
Figure 4. The predicted electric energy consumption and the actual demand. We show the prediction results for (a) 15, (b) 30, (c) 45, and (d) 60 min.
Energies 12 00739 g004
Figure 5. The training and test loss values for each epoch. We show the mean squared error of the model that predicts electric energy demand for (a) 15, (b) 30, (c) 45, and (d) 60 min.
Figure 5. The training and test loss values for each epoch. We show the mean squared error of the model that predicts electric energy demand for (a) 15, (b) 30, (c) 45, and (d) 60 min.
Energies 12 00739 g005
Figure 6. The results of MSE of models. We show the MSE results for (a) 15, (b) 30, (c) 45, and (d) 60 min.
Figure 6. The results of MSE of models. We show the MSE results for (a) 15, (b) 30, (c) 45, and (d) 60 min.
Energies 12 00739 g006
Figure 7. Visualization of states for monthly electricity demands. We describe the cluster of states as red circles on a monthly basis.
Figure 7. Visualization of states for monthly electricity demands. We describe the cluster of states as red circles on a monthly basis.
Energies 12 00739 g007
Figure 8. Examples of state transition from state in APR to state in AUG or OCT.
Figure 8. Examples of state transition from state in APR to state in AUG or OCT.
Energies 12 00739 g008
Table 1. Results of statistical analysis of electric energy demand by month, date, and hour.
Table 1. Results of statistical analysis of electric energy demand by month, date, and hour.
Statistical AnalysisMonthDateHour
p-value (t-test)0.0050.1820.011
p-value (ANOVA)0.0006.986 × 10 56 0.000
Table 2. The summary of related works.
Table 2. The summary of related works.
CategoryAuthorMethodDescription
Statistical modelMunz et al. (2007)K-means clusteringCalculate the center value of the cluster and classify the time series into regular and irregular trend
Kandananond (2011)ARIMAPresent different forecasting methods to predict electricity consumption
Cauwer et al. (2015)Statistical model and its physical principlesApply statistical method of multiple linear regression to real-world trip and energy consumption data
Machine learning modelDong et al. (2005)Support vector machinePredict building energy demand in the tropical region using SVM
Gonzalez and Zamarreno (2005)Feedback neural networkPresent a simple model to predict electric load in buildings
Ekici and Aksoy (2009)Artificial neural networkForecast buildings energy needs benefitting from orientation, insulation thickness and transparency ratio using ANN
Li et al. (2010)Support vector machineForecast annual energy demand using the building’s heat transfer coefficient
Xuemei et al. (2010)Fuzzy support vector machine, Fuzzy c-means clusteringPresent a novel short-term cooling load forecasting with conjunctive use of fuzzy C-mean clustering to define state and fuzzy SVM for prediction
Ma (2010)Linear regressionPropose model based on linear regression that predicts large-scale public building energy demand
Deep learning modelAhmad et al. (2017)Deep neural networkForecast energy demand with information of climate, date, and building usage rate
Lee et al. (2017)Recurrent neural networkMeasure the environmental consumption level for each country with a pro-environmental consumption index using big data queries
Li et al. (2017)AutoencoderExtract the building energy demand and predict future energy consumption
Table 3. The summary of the data used in this paper.
Table 3. The summary of the data used in this paper.
AttributeDateGAP (kW)GRP (kW)Voltage (V)GI (A)S1S2S3
Average-1.0890.124240.8444.6181.1171.2896.453
Std. Dev-1.0550.1133.2394.4356.1395.7948.436
Max26/11/201011.1221.390254.15048.40088.00080.00031.000
Min16/12/20060.0760.000223.2000.2000.0000.0000.000
Table 4. The numerical results of experiments. Our model outperforms the other models in most metrics.
Table 4. The numerical results of experiments. Our model outperforms the other models in most metrics.
MetricLRDTRFMLPLSTMSLSTMLi et al.Ours
15 min
MSE0.30770.30920.29152.09920.28530.27040.25310.2113
MAE0.28780.28410.27581.05800.32060.31190.31910.2517
MRE0.39250.39100.37192.23270.50210.48860.57060.3625
30 min
MSE0.42010.41250.39920.73150.40840.39640.32560.2937
MAE0.36570.35850.35110.63180.42320.41010.35810.3315
MRE0.53430.51650.50791.23250.71310.67780.60040.5399
45 min
MSE0.49820.48790.47570.74620.47200.45810.38080.3450
MAE0.42100.40970.40330.64420.47420.45590.41200.3700
MRE0.64620.60770.60121.26130.82740.76680.74410.6612
60 min
MSE0.55730.54570.53220.76210.49970.49450.41590.3840
MAE0.46260.44700.44090.62600.49460.49240.44040.3953
MRE0.73560.68280.67621.21610.88710.89160.81290.6478
Table 5. The results of experiments on state transition. Each column shows the month of the input electric energy demand and each row shows an output month after state transition. The ground truth (G.T.) is the average electric energy consumption for each month.
Table 5. The results of experiments on state transition. Each column shows the month of the input electric energy demand and each row shows an output month after state transition. The ground truth (G.T.) is the average electric energy consumption for each month.
JANFEBMARAPRMAYJUNJULAUGSEPOCTNOVDECG.T.
JAN-1.3331.3651.3911.3741.3391.3131.3141.3291.3411.3451.3791.389
FEB1.195-1.1571.1731.1581.1261.0831.0751.0841.0931.1151.1831.116
MAR1.1631.070-1.1411.1231.0931.0381.0491.0501.0521.0761.1471.131
APR1.1891.1041.158-1.1511.1221.0711.0681.0741.0841.1081.1741.176
MAY1.1991.1231.1601.184-1.1341.0901.0671.0751.0941.1201.1891.159
JUN1.3001.2421.2691.2871.272-1.2291.2081.2221.2341.2431.2891.244
JUL0.9030.8180.8230.8170.8280.806-0.7630.7640.7700.7980.8930.786
AUG0.6940.6010.5840.5610.5850.5760.537-0.5170.5480.5750.6870.511
SEP0.7960.6990.6950.6770.6970.6840.6370.620-0.6470.6790.7910.623
OCT0.9170.8290.8360.8280.8350.8180.7660.7500.754-0.8180.9200.779
NOV1.0200.9280.9530.9560.9500.9290.8660.8520.8540.876-1.0220.923
DEC1.2751.2021.2481.2831.2481.2211.1751.1501.1621.1871.208-1.265

Share and Cite

MDPI and ACS Style

Kim, J.-Y.; Cho, S.-B. Electric Energy Consumption Prediction by Deep Learning with State Explainable Autoencoder. Energies 2019, 12, 739. https://0-doi-org.brum.beds.ac.uk/10.3390/en12040739

AMA Style

Kim J-Y, Cho S-B. Electric Energy Consumption Prediction by Deep Learning with State Explainable Autoencoder. Energies. 2019; 12(4):739. https://0-doi-org.brum.beds.ac.uk/10.3390/en12040739

Chicago/Turabian Style

Kim, Jin-Young, and Sung-Bae Cho. 2019. "Electric Energy Consumption Prediction by Deep Learning with State Explainable Autoencoder" Energies 12, no. 4: 739. https://0-doi-org.brum.beds.ac.uk/10.3390/en12040739

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop