A Hierarchical Self-Organizing Map Model in Short-Term ... - CiteSeerX

18 downloads 43 Views 161KB Size Report
A Hierarchical Self-Organizing Map Model in Short-Term Load ... With power systems growth and the increase in their complexity, many factors have become ...
A Hierarchical Self-Organizing Map Model in Short-Term Load Forecasting

Ot´ avio A. S. Carpinteiro & Alexandre P. Alves da Silva Instituto de Engenharia El´etrica Escola Federal de Engenharia de Itajub´ a Av. BPS 1303, Itajub´ a, MG, 37500–903, Brazil e-mails: {otavio,alex}@iee.efei.br

Abstract This paper proposes a novel neural model to the problem of short-term load forecasting. The neural model is made up of two self-organizing map nets — one on top of the other. It has been successfully applied to domains in which the context information given by former events plays a primary role. The model was trained and assessed on load data extracted from a Brazilian electric utility. It was required to predict once every hour the electric load during the next 24 hours. The paper presents the results, and evaluates them.

Key words: short-term load forecasting, neural networks, self-organizing map.

1

Introduction

With power systems growth and the increase in their complexity, many factors have become influential to the electric power generation and consumption (load management, energy exchange, spot pricing, independent power producers, non-conventional energy, generation units, etc.); therefore, the forecasting process has become even more complex, and more accurate forecasts are needed. The relationship between the load and its exogenous factors is complex and non-linear, making it quite difficult to model through conventional techniques, such as time series and linear regression analysis. Besides not giving the required precision, most of the traditional techniques are not robust enough. They fail to give accurate forecasts when quick weather changes occur. Other problems include noise immunity, portability and maintenance [1]. Neural networks (NNs) have succeeded in several power system problems, such as: planning; control; analysis; protection; design; load forecasting; security analysis; and fault diagnosis. The last three are the most popular [2]. The NN ability in mapping complex non-linear relationships is responsible for the growing number of its application to the short-term load forecasting (STLF) [3, 4, 5, 6]. Several electric utilities over the world have been applying NNs for load forecasting in an experimental or operational basis [1, 2, 4]. So far, the great majority of proposals on the application of NNs to STLF use the multilayer perceptron trained with error backpropagation. Besides the high computational burden for su-

pervised training, multilayer perceptrons do not have a good ability to detect data outside the domain of the training data. The Kohonen’s self-organizing map has been proposed to overcome these shortcomings [7]. However, the Kohonen’s self-organizing map does not work appropriately for regression problems, because it does not preserve the functionality of the regression surface [8]. This paper introduces an extension of the original self-organizing map to STLF. It has been applied to load data extracted from a Brazilian electric utility. In the next sections, we shall introduce and detail the model, the experiment, evaluate the results, and draw some conclusions.

2

Representation for the Sequences

The input data in the experiment consists of sequences of load data. Eight neural input units are used in the representation. The first unit represents the load at the current hour. The second, the load at the hour immediately before. The third, the load two hours before. The fourth, fifth, and sixth units represent respectively the load at twenty-four hours behind, at one week behind, and at one week and twenty-four hours behind the hour whose load is to be predicted. The seventh and eighth units represent a trigonometric coding for the hour to be forecast, i.e., sin(2π.hour/24) and cos(2π.hour/24). Each unit receives real values. The load data is normalized, varying therefore, from zero to one.

3

The Model

The model is made up of two self-organizing maps (SOMs), as shown in Figure 1. Its features, performance, and potential are better evaluated in [9, 10]. Top

Map

SOM

Time Integrator Λ

Bottom SOM

Map Time Integrator V(t )

Figure 1: The model The input to the model is a sequence in time of m-dimensional vectors, S1 = V(1), V(2), . . . , V(t), . . . , V(z), where the components of each vector are non-negative real values. The sequence is presented to the input layer of the bottom SOM, one vector at a time. The input layer has m units, one for each component of the input vector V(t), and a time integrator. The activation X(t) of the units in the input layer is given by

X(t) = V(t) + δ1 X(t − 1)

(1)

where δ1 ∈ (0, 1) is the decay rate. For each input vector X(t), the winning unit i∗ (t) in the map1 is the unit which has the smallest distance Ψ(i, t). For each output unit i, Ψ(i, t) is given by the Euclidean distance between the input vector X(t) and the unit’s weight vector Wi . Each output unit i in the neighbourhood N ∗ (t) of the winning unit i∗ (t) has its weight Wi updated by

Wi (t + 1) = Wi (t) + αΥ(i)[X(t) − Wi (t)]

(2)

where α ∈ (0, 1) is the learning rate. Υ(i) is the neighbourhood interaction function [11], a Gaussian type function, and is given by

Υ(i) = κ1 + κ2 e−

κ3 [Φ(i,i∗ (t))] 2 2σ2

(3)

where κ1 , κ2 , and κ3 are constants, σ is the radius of the neighbourhood N ∗ (t), and Φ(i, i∗ (t)) is the distance in the map between the unit i and the winning unit i∗ (t). The distance Φ(i0 , i00 ) between any two units i0 and i00 in the map is calculated according to the maximum norm, Φ(i0 , i00 ) = max {|l0 − l00 | , |c0 − c00 |}

(4)

where (l0 , c0 ) and (l00 , c00 ) are the coordinates of the units i0 and i00 respectively in the map. The input to the top SOM is determined by the distances Φ(i, i∗ (t)) of the n units in the map of the bottom SOM. The input is thus a sequence in time of n-dimensional vectors, S2 = Λ(Φ(i, i∗ (1))), Λ(Φ(i, i∗ (2))), . . . , Λ(Φ(i, i∗ (t))), . . . , Λ(Φ(i, i∗ (z))), where Λ is a n-dimensional transfer function on a n-dimensional space domain. Λ is defined as

Λ(Φ(i, i∗ (t))) =



1 − κΦ(i, i∗ (t)) 0

if i ∈ N ∗ (t) otherwise

(5)

where κ is a constant, and N ∗ (t) is a neighbourhood of the winning unit. The sequence S2 is then presented to the input layer of the top SOM, one vector at a time. The input layer has n units, one for each component of the input vector Λ(Φ(i, i∗ (t))), and a time integrator. The activation X(t) of the units in the input layer is thus given by X(t) = Λ(Φ(i, i∗ (t))) + δ2 X(t − 1) 1 Also

known as array, grid, or output layer.

(6)

where δ2 ∈ (0, 1) is the decay rate. The dynamics of the top SOM is identical to that of the bottom SOM.

4

The Experiment on STLF

The experiment was on the prediction, once every hour, of electric load during the next 24 hours. The input data consisted of load data extracted from a Brazilian utility. The data was coded according to the representation presented in Section 2. The training of the two SOMs of the model took place in two phases — coarse-mapping and finetuning. In the coarse-mapping phase, the learning rate and the radius of the neighbourhood were reduced linearly whereas in the fine-tuning phase, they were kept constant. The bottom and top SOMs were tested and trained respectively with map sizes of 15×15 in 700 epochs, and 18×18 in 850 epochs. The initial weights were given randomly to both SOMs. The experiment comprised seven studies. In each study, one neural network was trained to hourly predict the load of a specific day of the week. Thus, the neural network of the first study was required to predict the load of Fridays. The network of the second study was to predict the load of Saturdays, and so on. Table 1 displays the percentage of the error as well as the mean error for each study. Figures 2– 8 plot the actual load, and the forecast load for the seven studies.

4000 3900 3800

Actual Load Forecast Load

Load (MW)

3700 3600 3500 3400 3300 3200 3100 3000 1537

1540

1543

1546

1549

1552

1555

Hour

Figure 2: Actual load and forecast load for the first study

1558

Table 1: Percentage of the error and mean error for each study    %  ' ! &  #     *%   *' ! (&  # % %  %% % %( -. /



 

 

   &&  !  ' " !$' " %  &  &(! " % %!  #  &* % )&  #   #  # ' ! % $# " % %( " % &*  &' % && *%  # % &'

  '  '*' % #! "+ &% % '!  %'  % % %( %% & ! ! % ! *' !! % '# (&)(& % ' % % &)   $' %  ! ' ! )&  %( (&) $  )&

!     $ #   )% ' !$' ! % *' % %% %! %  !  &(   %*'  !  $    &%  %# ' )% " % !$% " % #  %$& ",*' ", % ' !!



 



 

"  "  "+ % " &( " && " ' $! ",*' "# ! "#  "+ &' " # " ' "+ )% " !! " # # " %*' ", ! ", $% " $  &(! % #$%  # !(&  &%

 #  " $! " && "% ", &% "% # ", $ "  " &' " # " $%    '(  #  !  ! ",! " $' "  %#  %( !    % !$'

#  $!  (&) # % ! )& ! )% ' &* " ' % ! " ' &)   !   %   #$' ! )' '*%  !# (& % $ "+ !)& % &( ! &'  &(#  $  ##

  *% #)& ! $% ! ! ' ! !(& % % ", % *'  $%  &)  && % )& %(& *'  $  # "% $% ", '$&  ## " )' "% !# *%  !$'  )&

3600 3500 3400

Actual Load Forecast Load

Load (MW)

3300 3200 3100 3000 2900 2800 2700 2600 1561

1564

1567

1570

1573

1576

1579

1582

Hour

Figure 3: Actual load and forecast load for the second study

3300 3200 3100

Actual Load Forecast Load

Load (MW)

3000 2900 2800 2700 2600 2500 2400 2300 1585

1588

1591

1594

1597

1600

1603

1606

Hour

Figure 4: Actual load and forecast load for the third study

3600 3500 3400

Actual Load Forecast Load

Load (MW)

3300 3200 3100 3000 2900 2800 2700 2600 1609

1612

1615

1618

1621

1624

1627

Hour

Figure 5: Actual load and forecast load for the fourth study

1630

3600 3500 3400

Actual Load Forecast Load

Load (MW)

3300 3200 3100 3000 2900 2800 2700 2600 1633

1636

1639

1642

1645

1648

1651

1654

Hour

Figure 6: Actual load and forecast load for the fifth study

3500 3400 3300

Actual Load Forecast Load

Load (MW)

3200 3100 3000 2900 2800 2700 2600 2500 1657

1660

1663

1666

1669

1672

1675

Hour

Figure 7: Actual load and forecast load for the sixth study

1678

3500 3400 3300

Actual Load Forecast Load

Load (MW)

3200 3100 3000 2900 2800 2700 2600 2500 1681

1684

1687

1690

1693

1696

1699

1702

Hour

Figure 8: Actual load and forecast load for the seventh study The results are very promising, specially for several steps ahead forecasting. However, for a few steps ahead, the performance of the proposed model is poor. This behaviour confirmed our expectations. Medium decay rates extend the memory size for past events, and consequently, yield more accurate predictions on large horizons. As a side effect, however, in the predictions of the first hours of a day, medium decay rates led the model to take also into consideration the load from the last hours of the day immediately before, which in general shares little similarity with that of the day in question. More investigation on the effects of time integration is needed in order to produce a better adaptability. It is possible that one unit in the grid which has not responded to any load pattern during the training phase be requested to respond to a particular one during the prediction phase. The occurrence of this situation was rare, although it has produced highly undesirable effects. These effects may be seen, for example, after the points marked with arrows in figure 3. The neural network started producing very bad predictions, and it never recovered. A strategy must be planned to avoid this situation. More, continuous load data was employed as input in all studies. Yet, this data should have been refined to include just the pertinent data in each study. For instance, to hourly predict the load on a specific Monday, continuous load data including that of all days of the week was employed. Thus, the neural model was trained with irrelevant data once load patterns usually vary considerably from one day to the next. We believe that the refinement of the input data will improve the overall performance of the model.

5

Conclusion

A novel artificial neural model for sequence classification and prediction is presented. The model has a topology made up of two self-organizing map networks, one on top of the other. It encodes and manipulates context information effectively.

The results obtained have shown that the artificial neural model was able to perform efficiently the prediction of the electric load, particularly for larger forecasting horizons. We intend to do further research on the model, and study the effects of the time integrators on the predictions.

References [1] A. Khotanzad and et al. An adaptive modular artificial neural hourly load forecaster and its implementation at electric utilities. IEEE Trans. on Power Systems, 10(3):1716–1722, August 1995. [2] H. Mori. State-of-the-art overview on artificial neural networks in power systems. In M. ElSharkawi and D. Niebur, editors, A Tutorial Course on Artificial Neural Networks with Applications to Power Systems, chapter 6, pages 51–70. IEEE Power Engineering Society, 1996. [3] K. Liu and et al. Comparison of very short-term load forecasting techniques. IEEE Trans. on Power Systems, 11(2):877–882, May 1996. [4] A. Bakirtzis and et al. A neural network short-term load forecasting model for the greek power system. IEEE Trans. on Power Systems, 11(2):858–863, May 1996. [5] O. Mohammed and et al. Practical experiences with an adaptive neural network short-term load forecasting system. IEEE Trans. on Power Systems, 10(1):254–265, February 1995. [6] D. Park and et al. Electric load forecasting using an artificial neural network. IEEE Trans. on Power Systems, 6(2):442–449, May 1991. [7] T. Baumann and A. Germond. Application of the kohonen network to short-term load forecasting. In Proceedings of the Second International Forum on Application of Neural Networks to Power Systems, pages 407–412, Yokohama, Japan, April 1993. [8] V. Cherkassky and F. Mulier. Learning from Data. John Wiley & Sons, 1998. [9] O. A. S. Carpinteiro. A hierarchical self-organizing map model for pattern recognition. In L. Caloba and J. Barreto, editors, Proceedings of the Brazilian Congress on Artificial Neural Networks 97 (CBRN 97), pages 484–488, Florian´ opolis, SC, Brazil, 1997. UFSC. [10] O. A. S. Carpinteiro. A hierarchical self-organizing map model for sequence recognition. In L. Niklasson, M. Bod´en, and T. Ziemke, editors, Proceedings of the International Conference on Artificial Neural Networks 98, pages 815–820, Sk¨ ovde, Sweden, 1998. Springer-Verlag. [11] Z. Lo and B. Bavarian. Improved rate of convergence in Kohonen neural network. In Proceedings of the International Joint Conference on Neural Networks, volume 2, pages 201–206, July 8–12 1991.

Suggest Documents