The Genetic Algorithm Estimates for the Parameters of Order p Normal Distributions (1) Un Algoritmo Genetico per la Stima dei Parametri di Distribuzioni Normali di Ordine p Salvatore Vitrano Dipartimento di Sociologia, Universita’ degli Studi “La Sapienza” Via Salaria 113, I-00198 Roma, e-mail:
[email protected]
Roberto Baragona Dipartimento di Sociologia, Universita’ degli Studi “La Sapienza” Via Salaria 113, I-00198 Roma, e-mail:
[email protected] Riassunto: Le distribuzioni normali di ordine p, anche note come distribuzioni esponenziali potenziate, costituiscono una naturale generalizzazione della distribuzione normale. La stima dei parametri presenta delle difficoltà non trascurabili sia per quanto riguarda la distorsione che la variabilità. Gli algoritmi genetici, sperimentati con successo in molte applicazioni dell’inferenza statistica, possono costituire in questo senso una valida alternativa nell’ambito dei metodi di stima proposti. Sono qui considerate la stima del parametro di locazione per piccoli campioni, e dell’ordine p per campioni di numerosità moderata. Un esperimento di simulazione illustra i risultati e i vantaggi che possono derivare dall’uso degli algoritmi genetici e permette un confronto con alcuni metodi precedentemente introdotti e largamente sperimentati. Keywords: Exponential power functions, Genetic algorithms, Jackknife estimators, Maximum-likelihood, Normal distributions of order p.
1. Introduction th
Subbotin (1923), by generalizing the 2 Schiaparelli axiom, introduced a family of error distributions named Exponential Power Functions (EPF). Several formulations of the EPF were extensively analyzed in the literature, but here we refer to the model known as Normal Distribution of Order p, defined on the whole real axis and characterized by the location µ, the scale σp, and the shape parameter p (p>0). A family of unimodal symmetric curves from this density function arises with different shapes for different values of p. The Laplace (p=1), the Normal (p=2), and the Uniform (p → ∞ ) distributions arise as special cases. Let x=(x1, x2, …, xn) be a random sample drawn from this probability density function. The maximum-likelihood (ML) estimator of the parameter vector θ =(µ, σp, p) is any value θˆ which maximizes the likelihood function: n
Lθ (x) = {2 p1/pσp Γ(1+1/p)}–n exp{–(pσ pp ) −1 ∑ | x i i =1
(1)
- µ | p }.
(1)
The present paper is financially supported by the University of Rome “La Sapienza” and by M.U.R.S.T. (Italy) as a part of the grant “Non linear models and new computational methods in time series analysis and forecasting,“ 2000.
Properties and parameter estimation methods for the distributions in this class were extensively studied (see Chiodi, 2000, for a comprehensive review). We propose a genetic algorithm (GA) approach for estimating θ . Holland (1975) studied the evolution of a population of individuals in a given environment by means of a class of analytical models called GA. Each individual in the population, which has not to be given a strictly statistical meaning, needs to be characterized by a chromosome, i.e. a string of characters. Each character in the string possesses a specific meaning that may be decoded from both its value, called allele, and position, called locus. The population evolves through several generations towards the best adaptation to the environment. A basic assumption is that a function may be defined that maps the genetic pool (the set of all admissible strings) into the positive real axis. This function is called the fitness function, and it has to increase as soon as the adaptation to the environment increases. The special applications of the GA for function optimization assume the objective function as the fitness function, and the string owned by the best fit individual as the solution to the optimization problem. Chatterjee and Laudatto (1997) investigated using GA for statistical inference, and we closely followed the guidelines provided therein. The plan of the paper is as follows. The next section is devoted to the illustration of the GA we implemented for assessing estimates and standard errors. Comparison with alternative approaches by means of a simulation experiment is provided in Section 3. The estimation of the location parameter µ in case of small samples, for given order p, and the estimation of all three parameters, for rather large sample sizes, is considered.
2. Genetic algorithm implementation The main features of our design of the GA are the tournament selection for reproduction, and the use of three operators, crossover, inversion, and mutation. Unlike many GA applications, inversion proved to constitute a valid support to increase the speed of convergence to the solution. Further, coding the potential solutions, that are real numbers in a given interval (a,b), say, was done by using the following device. Let us choose a positive integer as the binary string length. Then, 2 different numbers x may be coded according to x = a + c (b – a) / (2 – 1),
(2)
where c represents the real value corresponding to the given binary string. As c may vary from 0 (the null string 000…000) through 2–1 (the all-one string 111…111), then (2) encodes a finite subset of real numbers in the interval (a,b), at equispaced intervals of length (b–a)/(2–1) each. As far as the GA parameters choice is concerned, for the present problem, we assumed the selection pressure ps=0.75, the crossover, mutation, and inversion probability pc=0.7, pi=0.65, and pm=0.01 respectively, and 50 as the number of generations. We assumed the population size rather small, s=50, as no substantial improvement was observed by taking larger values. The likelihood (1) provided the fitness function.
Table 1: Order-p Normal distribution parameter estimates, all parameters unknown µ p=1 n=50 n=100 p=1.25 n=50 n=100 p=1.50 n=50 n=100 p=1.75 n=50 n=100 p=2 n=50 n=100 p=2.25 n=50 n=100 p=2.5 n=50 n=100 p=2.75 n=50 n=100 p=3 n=50 n=100 p=3.25 n=50 n=100
Maximum likelihood p σp
µ
Genetic algorithm σp
p
3.0193
2.1429
1.2462
3.0267
2.1341
(.2868)
(.3198)
(.3390)
(.2796)
(.3171)
(.3279)
2.9781
2.0606
1.1144
3.0074
2.0403
1.1099
(.2004)
(.1917)
(.1602)
(.2169)
(.1988)
(.1762)
2.9765
2.0486
1.5072
3.0235
2.0528
1.4813
(.3578)
(.3436)
(.7016)
(.3176)
(.3237)
(.6211)
3.0447
2.0468
1.3691
3.0066
2.0593
1.3774
(.2181)
(.2495)
(.2954)
(.2063)
(.2820)
(.3766)
2.9694
1.9805
1.7014
2.9812
2.0096
1.8719
(.3106)
(.3483)
(.7612)
(.3167)
(.3783)
(.9849)
2.9924
1.9949
1.5935
2.9922
2.0094
1.5980
(.1959)
(.2634)
(.3806)
(.2009)
(.2324)
(.4399)
3.0454
2.1079
2.2741
2.9633
2.0760
2.1561
(.3220)
(.4431)
(1.0201)
(.2962)
(.3908)
(.9725)
2.9570
1.9861
1.8479
2.9894
1.9606
1.8190
(.1987)
(.2150)
(.3964)
(.2111)
(.2353)
(.4464)
3.0064
2.0554
2.5580
3.0510
2.0858
2.5986
(.2908)
(.3666)
(1.1405)
(.2963)
(.3090)
(1.0214)
2.9938
2.0654
2.2843
3.0127
2.024
2.2229
(.2196)
(.2362)
(.7219)
(.1815)
(.2246)
(.6650)
3.0160
2.0807
3.0267
2.9586
2.0637
2.8037
(.2985)
(.3049)
(1.1726)
(.2848)
(.3257)
(1.1711)
3.0314
1.9904
2.4970
2.9944
2.0054
2.4778
(.2077)
(.2345)
(.6559)
(.1883)
(.2125)
(.7259)
2.9395
2.0682
3.2349
2.9985
1.9875
3.0510
(.2562)
(.2919)
(1.1717)
(.2670)
(.3054)
(1.2151)
3.0771
2.0807
2.9473
3.0098
2.0395
2.7787
(.1718)
(.2123)
(.8330)
(.1661)
(.2171)
(.7440)
2.9752
2.0459
3.2150
3.0213
1.9974
3.2069
(.2416)
(.3026)
(1.1268)
(.2671)
(.3146)
(1.1257)
2.9778
2.0149
3.0505
2.9904
1.9498
2.8491
(.1846)
(.2213)
(.8972)
(.1846)
(.1966)
(.7361)
2.9970
2.0584
3.6761
3.0009
2.0152
3.5714
(.2448)
(.2661)
(1.1710)
(.2509)
(.3022)
(1.2258)
3.0188
2.0110
3.4036
3.0051
2.0365
3.3715
(.1733)
(.2124)
(.9240)
(.1624)
(.1970)
(.8667)
3.0054
1.9841
4.0412
3.0103
1.9782
3.7364
(.2199)
(.2248)
(1.0605)
(.2260)
(.2782)
(1.2226)
2.9912
2.0242
3.6140
3.0063
2.0174
3.5910
(0.1609)
(.1932)
(.9203)
(.1639)
(.2161)
(.9309)
Note: the estimated standard errors are given in parentheses.
K
1.2250 10 5 5 6 2 1 2
1
3. Simulation and results To compare the parameter estimates obtained by the GA with the ML estimates obtained by the common search numerical algorithms, two distinct Monte Carlo simulations were designed. The numerical inspection of the sample mean and standard deviation of the parameter estimates was made by drawing 100 samples using of size n= 10,20,50, when p was supposed known, and n=50, 100, when p was unknown, from a Normal distribution of order p (µ=3; σp=2) with p = 1-3.25 (0.25), through a generator algorithm (Chiodi, 1986). Some simulation results are given in Table 1. The number K of samples, for which the ML procedure has not reached any optimum point of (1), was also reported along with the parameter estimates by the two approaches. When p>2, a less bias was observed using the GA than the ML approach. Opposite results occur if p≤2. In this case, however, some samples were discarded by the ML procedure. Similar findings were obtained, except some particular cases, for the estimates of the two other parameters. When p was assumed known, the GA and two different scale parameter estimators (Chiodi, 1988; La Tona, 1994) were considered. The estimators gave results that were found to differ less than 1/1000, with best results by using the GA, when p>2. Finally, in both types of Monte Carlo simulations (p known and unknown), the standard deviations of the estimates from the two approaches were obtained that had the same order of magnitude.
References Chatterjee, S., Laudatto, M. (1997) Genetic algorithms in statistics: procedures and applications, Communications in Statistics - Simulation and Computation, 26(4), 16171630. Chiodi, M. (1986) Procedures for generating pseudo-random numbers from a normal distribution of order p (p>1), Statistica Applicata, 1, 7-26. Chiodi, M. (1988) Sulle distribuzioni di campionamento delle stime di massima verosimiglianza dei parametri delle curve normali di ordine p, Istituto di Statistica Università di Palermo. Chiodi, M. (2000) Le curve normali di ordine p nell’ambito delle distribuzioni di errori accidentali: una rassegna dei risultati e problemi aperti per il caso univariato e per quello multivariato, in: Atti della XL Riunione Scientifica della SIS, ISTAT Centro Stampa, 5960. Extended and revised version available at http://statistica.economia.unipa.it/ Holland, J. H. (1975) Adaptation in Natural and Artificial Systems. University of Michigan Press (Second Edition: MIT Press, 1992) La Tona, L. (1994) Uno stimatore jackknife del parametro di scala di una curva normale di norma p, Metron, 52, 181-188. Subbotin, M. T. (1923) On the law of frequency of errors, Matematicheskii Sbornik, 31, 296-300.