Research Article  Open Access
Zhisheng Zhang, Wenjie Gong, "ShortTerm Load Forecasting Model Based on Quantum Elman Neural Networks", Mathematical Problems in Engineering, vol. 2016, Article ID 7910971, 8 pages, 2016. https://doi.org/10.1155/2016/7910971
ShortTerm Load Forecasting Model Based on Quantum Elman Neural Networks
Abstract
Shortterm load forecasting model based on quantum Elman neural networks was constructed in this paper. The quantum computation and Elman feedback mechanism were integrated into quantum Elman neural networks. Quantum computation can effectively improve the approximation capability and the information processing ability of the neural networks. Quantum Elman neural networks have not only the feedforward connection but also the feedback connection. The feedback connection between the hidden nodes and the context nodes belongs to the state feedback in the internal system, which has formed specific dynamic memory performance. Phase space reconstruction theory is the theoretical basis of constructing the forecasting model. The training samples are formed by means of nearest neighbor approach. Through the example simulation, the testing results show that the model based on quantum Elman neural networks is better than the model based on the quantum feedforward neural network, the model based on the conventional Elman neural network, and the model based on the conventional feedforward neural network. So the proposed model can effectively improve the prediction accuracy. The research in the paper makes a theoretical foundation for the practical engineering application of the shortterm load forecasting model based on quantum Elman neural networks.
1. Introduction
Shortterm load forecasting (STLF) is the basis for the normal and safe operation of power system. Since the introduction of competition mechanism in power market, the power company has put forward higher requirements on the accuracy and rapidity for load forecasting [1–3]. Many classical methods were applied to the study of STLF model, such as ARIMA model and ARMAX model [4, 5]. However, the ability of these classical methods to deal with the nonlinear time series is poor. With the introduction of artificial intelligence technology to the study of STLF model, it has opened up many new research methods such as neural network and expert system [6, 7]. Because of the strong nonlinear time series processing ability, neural network is widely used in the study of STLF.
The process of human brain information processing may be related to the quantum phenomenon, and quantum mechanical effects may exist in the brain, which had been shown by some research results [8]. The quantum systems have the similar dynamic characteristics with the biological neural networks. The fusion of artificial neural network and quantum theory can better simulate the process of human brain information processing. Quantum neural network can improve the approximation ability and information processing efficiency of artificial neural network, which had been used for many applications [9–13].
The STLF model based on quantum Elman neural networks (QENN) is constructed in this paper. QENN is composed of the quantum neurons. The quantum neuron model consists of the weighting part, the aggregating part, the activating part, and the excitation part. Different from the traditional multilayer feedforward neural network, QENN has not only the feedforward connection but also the Elman feedback connection. The Elman feedback connection between the hidden nodes and the context nodes belongs to the state feedback in the internal system, which has formed specific dynamic memory performance [14].
Some studies had shown that the load time series is nonlinear and chaotic [15]. Phase space reconstruction theory (PSRT) is the important analysis method for research on chaotic time series. The training samples are formed by means of nearest neighbor approach (KNNA) based on PSRT.
Through the actual example simulation, it is proved that the proposed model can effectively improve the prediction accuracy and has adaptability to different load time series.
2. Quantum Elman Neural Networks
QENN is composed of quantum neurons, and quantum neurons are the basic elements of building QENN.
2.1. Quantum Neuron Model
2.1.1. Qubit
In the classical computation, the binary numbers “0” and “1” are used to represent the information, and they are usually called bits. In the quantum computation, and are used to represent the two basic states of microscopic particles, and they are usually called qubits. The difference between qubits and bits is that, in addition to the state of and , the state of qubits can be a linear combination of states, which is usually called the superposition.
In quantum computation, a qubit state can be expressed aswhere and express the basic states of microscopic particles. and are the probability amplitudes, and they should meet the following relations:
In order to further understand the superposition of the qubits, an electron is used to illustrate. An electron can be in the ground state of low energy level, and it can also be in the excitation state of high energy level. According to the superposition principle, the electron can be in the state of linear combination of these two states, that is, the superposition state . It is shown in Figure 1.
If the electron is in the superposition state , the electron can select the ground state according to the probability and select the excitation state according to the probability . It is shown in Figure 2.
2.1.2. Quantum Neuron Model
The quantum neuron model adopted in the paper consists of the weighting part, the aggregating part, the activating part, and the excitation part. The weighting part is the simulation of the bond strength between synapses of nerve cells. The aggregating part is the simulation of the spatial and temporal integration of stimuli received by multiple dendrites. The activating part is the simulation of the interaction between the current activation values and membrane potential change of nerve cells. The excitation part is the simulation of the nonlinear characteristics of nerve cells, such as excitation, inhibition, fatigue, refractory period, and threshold. According to [16], the quantum behavior of neural cells is mainly reflected in the neurotransmitter transmission of the cells. Transmitter release from nerve terminals in the form of small packages, with multiple molecules for biological neurons, and the small package is called quantum. A small package or quantum describes the smallest unit of transmitter secretion. In general, the quantum content of a synaptic response may be different, but the number of molecules in a quantum is relatively certain; that is, the quantum size is relatively determined. Since 1990s, some research results show that the process of human brain information processing may be related to the quantum phenomenon, and quantum mechanical effects may exist in the brain, and the quantum systems have the similar dynamic characteristics with the biological neural networks. So combining artificial neural network with quantum theory can better simulate the process of human brain information processing.
The quantum neuron model is shown in Figure 3. Considering the actual prediction problem, the input vector is realvalued vector, and the output is the real number.
The weight and activation values are qubits. is the weight vector, and is the weight between the input and the quantum neuron. can be described as
is the activation value, which can be described as
The activation values are the important parameters for the activating part. The simulation of the interaction between the current activation values and membrane potential change of nerve cells is the core of the activating part. The activation values affect the accuracy of prediction. The parameters can be optimized by genetic algorithm.
In Figure 1, represents the aggregation operator, and represents the activation function. is the excitation function, which adopts the Sigmoid function, and it can be described as
The relationship between input and output can be expressed aswhere is the threshold.
2.2. Quantum Elman Neural Networks Model
Considering Figure 4, QENN has four layers which are the input layer, the hidden layer, the context layer, and the output layer. The hidden layer and the output layer are constructed by quantum neurons.
For QENN, each node in the hidden layer has a corresponding node connected to it in the context layer.
The input nodes and the hidden nodes, the hidden nodes and the output nodes, and the context nodes and the hidden nodes are connected by adjustable weights.
Compared with the feedforward network structure, QENN has the context layer. The role of the nodes in the context layer is to store the internal state of neural network. They are used to store the state of the hidden nodes at the current time, and the state will be passed to the hidden nodes at the next time. The connection between the hidden nodes and the context nodes belongs to the state feedback in the internal system, which has formed specific dynamic memory performance.
QENN can not only store the information in the current input data but also store some historical information. In the training process, it can dynamically backtrack to the input at the past time related to the expected output at the current time.
, are the number of the input nodes and the hidden nodes, respectively. The number of the context nodes is equal to the number of the hidden nodes. There is one output node. The number of iterations is denoted as . is the weight between th input node and th hidden node, and it can be described as
is the weight between th context node and th hidden node, and it can be described as
is the weight between th hidden node and th output node, and it can be described as
is the input value of th external input neuron at time .
is the internal state of th context neuron at time , which is the output of th hidden neuron at time , and it can be expressed aswhere is the output of th hidden neuron at time .
is the output of th context neuron at time , and it can be expressed as
is the internal state of th hidden neuron at time , which can be expressed aswhere is the activation value of th hidden neuron, which can be described as
is the output of th hidden neuron at time , which can be expressed as
is the internal state of the output neuron at time , and it can be expressed aswhere is the activation value of th output neuron, which can be described as
is the output of the output neuron at time , which can be expressed as
2.3. Multiplicity of Convergent Solution for QENN
According to (10)–(17), can also be expressed as
If and are the convergent solution of the sequence and , and which satisfy are the convergent solution of the sequence and .
This shows that many groups of and are the convergent solution of the sequence and . It only requires the argument of and to satisfy a certain relation, and the specific value of the argument is not required. This ensures the multiplicity of convergent solution for QENN, which is the basis of improving the approximation capability of the neural networks and making its performance enhanced in the application to a shortterm load prediction of QENN.
2.4. Parameters Optimization of QENN by Genetic Algorithm
Genetic algorithm is used to adjust the parameters of the QENN, which include
The chromosome can be constituted by these parameters, which can be expressed as
Genetic algorithm uses the roulette method as selecting operator. The crossover operator adopts the nonuniform arithmetic crossover method, and the mutation operator uses the noncoherence mutation method. The fitness function mainly considers the accuracy of the output of the forecasting model; that is, . is the mean square deviation between the desired output and the actual output of the forecasting model.
The process is shown in Figure 5.
2.4.1. Nonuniform Arithmetic Crossover Operator
The operation object of arithmetic crossover operator is chromosome , which is encoded by floating point numbers.
If two chromosome individuals are and , the two new individuals generated by the crossover operation can be expressed as
For the nonuniform arithmetic crossover operator, is the random number, and its value range is from −0.5 to 1.5.
2.4.2. Noncoherence Mutation Operator
Supposing the chromosome individual is a parent solution, the component is selected to mutate, and its value range is . The mutated chromosome individual is .
can be expressed as where is the current evolution generation, and can be expressed aswhere is a random number and its value range is from 0 to 1. is the maximal generation. is a parameter for the nonuniform arithmetic crossover operator, and its value is 2.5.
3. PSRT and KNNA
3.1. Phase Space Reconstruction Theory
The chaotic time series is usually analyzed by PSRT. The embedding theorem of Takens and Saner et al. is the basis of the prediction theory to chaotic time series.
To chaotic time series , the phase space can be reconstructed by PSRT, which can be described aswhere is the embedding dimension and it can be acquired by the GP method [17]. is the delay time, and it can be acquired by multipleautocorrelation method [18]. In the dimension phase space, the evolution of the chaotic system can be described by the evolution track of phase points.
According to the GP method, the embedding dimension is 7. By multipleautocorrelation method, the delay time is 1 for the real example simulation in this paper.
PSRT provides a theoretical basis for the construction of the QENN model. For the model, the number of input nodes is decided by the embedding dimension according to PSRT.
3.2. Nearest Neighbor Approach
The nearest neighbor approach is usually used in the forecasting research of chaotic time series [19], which is based on the Euclid distance. The similarity of two phase points is computed by the Euclid distance.
In the reconstructed phase space with dimension, the forecasting phase point is recorded as , and denotes th neighbor phase point. is the Euclid distance between the forecasting phase points and th neighbor phase point, which is expressed as
According to (24), the nearest neighbor phase points set can be formed, which is the training samples set of forecasting phase points for the QENN model.
4. Real Example Simulation
4.1. Four STLF Models for Real Example Simulation
Four STLF models are constructed, which are modelI based on the QENN, modelII based on the quantum feedforward neural network, modelIII based on the conventional Elman neural network, and modelIV based on the conventional feedforward neural network.
The input nodes number of four STLF models for actual simulation load system is 7, which is acquired by the GP method. The delay time is 1, which is acquired by multipleautocorrelation method.
The number of the neighbor phase points affects the prediction accuracy and calculation of the forecasting model. If the number of the neighbor phase points is too much, it will not only increase the amount of computation but also increase the number of false neighbor phase points, which has a great impact on the prediction performance. If the number of the neighbor phase points is too small, the effective information is not fully utilized, which also affects the prediction performance. In general, the relationship between the number of the neighbor phase points and the embedding dimension satisfies the following conditions: . Based on multiple validation tests, the number of the neighbor phase points is 10 for the actual simulation load system in this paper.
For modelI and modelIII, the number of the input nodes is 7, the number of the hidden nodes is 10, the number of the context nodes is 10, and the number of the output points is 1.
For modelII and modelIV, the number of the input nodes is 7, the number of the hidden nodes is 10, and the number of the output points is 1.
The maximum permissible error of the four STLF models is . For modelI, modelII, modelIII, and modelIV, genetic algorithm was used to adjusted the parameters of QENN, the quantum feedforward neural network, and the conventional Elman neural network. The fitness function mainly considers the accuracy of the output of the forecasting model. The crossover probability is 0.90, and the variation probability is 0.10.
4.2. Simulation Results
Four models are used to predict the actual load. The daily load forecasting errors and comparison of four models for workday and weekend are shown in Figures 6 and 7, respectively.
The mean absolute percentage error and the maximum relative error of four STLF models for workday and weekend are described in Table 1. The mean absolute percentage error can be calculated aswhere is the number of the forecasting points, is the forecasting load, and is the actual load.

The maximum relative error is the maximum value of relative error of all predicted points. The relative error can be calculated as
The error performance of the modelIV is the worst in the all models, because the conventional feedforward neural network cannot effectively characterize the chaotic dynamic behavior of the load system. The error performance of modelII is better than that of modelIV, because quantum computation can enhance information processing capability and generalization ability of quantum feedforward neural networks. The error performance of modelIII is better than that of modelIV, because the conventional Elman neural network has formed specific dynamic memory performance. The forecasting performance of modelI is the best, which is the fusion of quantum computation and feedback mechanism.
From Table 1 it can be seen that the forecasting error is larger in weekend. In weekend, the load mode is more diverse and complex, which leads to the forecasting error is slightly larger than that of workday.
In order to verify how sensitive are the simulation results on and , the forecasting errors of modelI with different are described in Table 2, and the forecasting errors of modelI with different are described in Table 3.


From Tables 2 and 3, it can be seen that the forecasting error is related to and . For the actual simulation load system in this paper, and are appropriate.
In order to further verify the performance of modelI, the real example are used to test in a week. The mean absolute percentage error and the maximum relative error in the week are described in Table 4. It is shown that the proposed modelI has good precision performance and steadiness.

5. Conclusions
This paper constructs a STLF model based on QENN. The quantum computation and Elman feedback mechanism were integrated into QENN. Quantum computation can effectively improve the approximation capability and the information processing ability of the neural networks. QENN has not only the feedforward connection but also the feedback connection. Phase space reconstruction theory is the theoretical basis of constructing the forecasting model. The training samples are formed by means of nearest neighbor approach. Through the example simulation, the testing results show that proposed model can effectively improve the prediction accuracy. The research in the paper makes a theoretical foundation for the practical engineering application of the STLF model based on QENN. The model based on QENN can also be used in the power prediction for new energy, such as wind power prediction and photovoltaic power prediction.
Competing Interests
The authors declare that they have no competing interests.
Acknowledgments
This work was supported by the project Supported by National Nature Science Foundation of China (51477078).
References
 E. Paparoditis and T. Sapatinas, “Shortterm load forecasting: the similar shape functional timeseries predictor,” IEEE Transactions on Power Systems, vol. 28, no. 4, pp. 3818–3825, 2013. View at: Publisher Site  Google Scholar
 S. Skolthanarat, U. Lewlomphaisarl, and K. Tungpimolrut, “Shortterm load forecasting algorithm and optimization in smart grid operations and planning,” in Proceedings of the 2nd Annual IEEE Conference on Technologies for Sustainability (SusTech '14), pp. 165–171, IEEE, Portland, Ore, USA, July 2014. View at: Publisher Site  Google Scholar
 E. Ceperic, V. Ceperic, and A. Baric, “A strategy for shortterm load forecasting by support vector regression machines,” IEEE Transactions on Power Systems, vol. 28, no. 4, pp. 4356–4364, 2013. View at: Publisher Site  Google Scholar
 A. K. Fard and M.R. AkbariZadeh, “A hybrid method based on wavelet, ANN and ARIMA model for shortterm load forecasting,” Journal of Experimental and Theoretical Artificial Intelligence, vol. 26, no. 2, pp. 167–182, 2014. View at: Publisher Site  Google Scholar
 H. K. Alfares and M. Nazeeruddin, “Electric load forecasting: literature survey and classification of methods,” International Journal of Systems Science, vol. 33, no. 1, pp. 23–34, 2002. View at: Publisher Site  Google Scholar
 Z. Xiao, S.J. Ye, B. Zhong, and C.X. Sun, “BP neural network with rough set for short term load forecasting,” Expert Systems with Applications, vol. 36, no. 1, pp. 273–279, 2009. View at: Publisher Site  Google Scholar
 K.H. Kim, J.K. Park, K.J. Hwang, S.H. Kim, H.G. Han, and S.H. Kang, “Implementation of shortterm load forecasting expert systems in a real environment,” International Journal of Engineering Intelligent Systems for Electrical Engineering and Communications, vol. 8, no. 3, pp. 139–144, 2000. View at: Google Scholar
 S. C. Kak, “On quantum neural computing,” Information Sciences, vol. 83, no. 34, pp. 143–160, 1995. View at: Publisher Site  Google Scholar
 V. Gandhi, G. Prasad, D. Coyle, L. Behera, and T. M. McGinnity, “Quantum neural networkbased EEG filtering for a braincomputer interface,” IEEE Transactions on Neural Networks and Learning Systems, vol. 25, no. 2, pp. 278–288, 2014. View at: Publisher Site  Google Scholar
 K. Takahashi, Y. Shiotani, and M. Hashimoto, “Remarks on model reference selftuning PID controller using quantum neural network with qubit neurons,” in Proceedings of the International Conference on Soft Computing and Pattern Recognition (SoCPaR '13), pp. 253–257, Hanoi, Vietnam, December 2013. View at: Publisher Site  Google Scholar
 P. Li, Y. Chai, and Q. Xiong, “Quantized neural modeling: hybrid quantized architecture in Elman networks,” Neural Processing Letters, vol. 37, no. 2, pp. 163–187, 2013. View at: Publisher Site  Google Scholar
 M. F. Ab Aziz and S. M. Hj Shamsuddin, “Quantum particle swarm optimization for Elman recurrent network,” in Proceedings of the 4th International Conference on Mathematical Modelling and Computer Simulation (AMS '10), pp. 133–137, Kota Kinabalu, Malaysia, May 2010. View at: Publisher Site  Google Scholar
 P.H. Li, Y. Chai, and Q.Y. Xiong, “Quantum gate Elman neural network and its quantized extended gradient backpropagation training algorithm,” Acta Automatica Sinica, vol. 39, no. 9, pp. 1511–1522, 2013. View at: Publisher Site  Google Scholar
 J. L. Elman, “Finding structure in time,” Cognitive Science, vol. 14, no. 2, pp. 179–211, 1990. View at: Publisher Site  Google Scholar
 A. C. Tsakoumis, P. Fessas, and V. M. Mladenov, “Application of chaotic time series for shortterm load prediction,” WSEAS Transactions on Systems, vol. 2, no. 3, pp. 517–523, 2003. View at: Google Scholar
 Z. Xu, G. Wu, and H. Wang, “Forecasting method for enhancing oil recovery in molecule deposition oildisplacement based on quantum neural network,” Acta Petrolei Sinica, vol. 29, no. 1, pp. 84–88, 2008. View at: Google Scholar
 P. Grassberger and I. Procaccia, “Measuring the strangeness of strange attractors,” Physica D: Nonlinear Phenomena, vol. 9, no. 12, pp. 189–208, 1983. View at: Publisher Site  Google Scholar  Zentralblatt MATH
 J. Y. Lin, Y. K. Wang, and Z. P. Huang, “Selection of proper timedelay in phase space reconstruction of speech signals,” Signal Processing, vol. 15, no. 3, pp. 220–225, 1999. View at: Google Scholar
 I. Drezga and S. Rahman, “Input variable selection for ANNbased shortterm load forecasting,” IEEE Transactions on Power Systems, vol. 13, no. 4, pp. 1238–1244, 1998. View at: Publisher Site  Google Scholar
Copyright
Copyright © 2016 Zhisheng Zhang and Wenjie Gong. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.