=Paper=
{{Paper
|id=Vol-1152/paper68
|storemode=property
|title=Method for Detection of Mixtures of Normal Distributions with Application to Vine Varieties
|pdfUrl=https://ceur-ws.org/Vol-1152/paper68.pdf
|volume=Vol-1152
|dblpUrl=https://dblp.org/rec/conf/haicta/OliveiraO11
}}
==Method for Detection of Mixtures of Normal Distributions with Application to Vine Varieties==
Method for detection of mixtures of normal distributions with application to vine varieties Amílcar Oliveira1, Teresa Oliveira2 1 Universidade Aberta, Department of Sciences and Technology Rua Fernão Lopes 9 2Dir, 1000 132 Lisbon Portugal, e-mail: aoliveira@univ-ab.pt 2 Universidade Aberta, Department of Sciences and Technology Rua Fernão Lopes 9 2Dir, 1000 132 Lisbon Portugal, e-mail: toliveira@univ-ab.pt Abstract. In this work we trait the problem of mixtures of normal distributions and methods for estimating the number of components as well as the parameters in a mixture. Also, we present a practical method for the detection of normal finite mixture distribution and respective model validation. Finally, we apply the exposed procedure to a sample of old grape-vine castes. Keywords: Normal distribution, mixtures, grape-vine caste. 1 Introduction Finite mixtures of distributions, and in particular mixtures of normal distributions, have been extensively used to model a wide variety of important practical situations, in which data can be considered from two or more populations mixed in varying populations. It is therefore evident interest in this subject attending the vast applications that have been developed by statisticians. We will emphasize some topics that have been successfully addressed in this area, which include among others the problem of identification of outliers, (Atkin & Tunnicliffe, 1980) (Wilson, 1980) or (Beckman & Cook, 1983), latent class models (Goodman, 1974), classification analysis (Symons, 1981) (Celeux, 1986) or (Bozdogan, 1992), investigating the robustness of certain statistics such as correlation coefficient sample studied (Srivastava & Lee 1984). In our work there will be some introductory remarks in the context of finite mixtures in order to create the enabling environment for a better understanding of this subject. We will continue with a statistical approach to key issues in the context of mixtures which will focus on the main methods of estimating parameters of a mixture and one of the most used algorithms in the identification of a finite mixture of distributions, i.e., the so-called EM algorithm (Dempster, et al., 1977). BBBBBBBBBBBBBBBBBBBBBBBBBBBBBBB &RS\ULJKWE\WKHSDSHU¶VDXWKRUV&RS\LQJSHUPLWWHGRQO\IRUSULYDWHDQGDFDGHPLFSXUSRVHV ,Q06DODPSDVLV$0DWRSRXORV HGV 3URFHHGLQJVRIWKH,QWHUQDWLRQDO&RQIHUHQFHRQ,QIRUPDWLRQ DQG&RPPXQLFDWLRQ7HFKQRORJLHV IRU6XVWDLQDEOH$JULSURGXFWLRQDQG(QYLURQPHQW +$,&7$ 6NLDWKRV6HSWHPEHU 771 2 Estimating the number of components in a mixture The problem of statistical analysis of finite mixtures can be divided into the following phases: i) Verification of identifiability of the mixture; ii) Estimation of the number of components / parameters estimation, iii) Testing the number of components iv) Validation of the model With a finite mixture density, f(x, r) function is identifiable if and only if å p f (x q ) = å p f (x q )Þ k = k Ù (" ) k k* j j * i * i * j =1,...,k ,$i=1,...,k* : p j = pi* Ù q j = qi* (1) j =1 j =1 That is to say that the mixture is identifiable if it admits only a single decomposition. (Teicher, 1963) deduced the necessary and sufficient conditions for identifiability and proved that the mixtures of normal distributions are identifiable. 3 Methods for the estimation of mixture parameters The problem of parameter estimation in mixtures, in the case of normal distributions, is one of the oldest problems in the statistical literature. It was first introduced by (Pearson, 1894) in an article "Contribution to the theory of evolution mathematical" and subsequently developed by (Quandt & Ramsey, 1978). It is still an open problem which attracts strong attention. Although at present the study of mixtures takes place in several areas by applying other methods, such as the method of maximum likelihood, the original method, the method of moments is still considered one of the best approximation methods in separating mixtures of normal distributions. It is useful even in the generation of initial estimates for the iterative resolution of maximum likelihood equations. In order to make the analysis of mixtures of distributions a computational problem more accessible, in the decades of the forties and fifties it was fostered the development of a high number of graphical techniques. A first step consisted in the detection of turning points of the curves (Harding, 1948) and (Cassie, 1954), making this method a somewhat subjective process. Later, more rigorous techniques have been suggested to determine the inflection points (Fowlkes, 1979) and (Bhattacharya, 1967); this author also suggests several methods to determine the proportions of the mixtures. These graphical techniques are not only to give a first estimation of parameters, they can be quite useful in an initial examination of data, since they have the advantage of running without a prior knowledge of the number of components of the mixture. They can play a role as an indicator of the number of components, since 772 this information is needed before applying any of the other methods described below. 3.1 Method of Moments This method is used on obtaining and solving a system of equations, often of the nonlinear type. The equations are obtained from the empirical equality of every moment (M r ) and their theoretical moment (mr ) , Mr = 1 n å (x j - x ) r , r = 1,...,m (2) n j =1 and mr = ò (x - m ) f (x ) dx , r = 1,..., m r (3) Where x represents the mean of a sample obtained from a population with probability density function f ( x ) , m is the mean value of random variable X , with the same probability function, m the number of moments needed to estimate all parameters. Let for each value of r obtaining an equation: 1 n å (x j - x ) r = ò (x - m )r f (x ) dx (4) n j =1 Note that with the currently available computational means this method becomes very advantageous. The disadvantage lies in the fact that it is not applicable to mixtures of distributions with a large number of components, as well as to the multidimensional case. 3.2 Method of Quandt and Ramsey This method proposed by (Quandt & Ramsey, 1978) is used in mixtures of two univariate components and makes use of the moment generating function E e ( ). tx The estimate for this function is given by: ( ) 1n å e n tx j Ê etx = (5) j =1 The method minimizes the sum of squared deviations between the empirical moment generating function and the theoretical moment generating function [ ( ) ( )] n 2 S (t ) = å Ê e j - E e j tx tx (6) j =1 773 where k represents the number of t values in a selected interval (a,b ) , with a < 0 and b > 0 . 3.3 Method of maximum likelihood Consider the sample values x1 ,..., xn , x j Î Âm , j= 1,...,n , a mixture of k density functions and consider the log-likelihood function of this sample, represented by én ù L(x q ) = å ln êå pi f (x j ai )ú n (7) j =1 ë j =1 û where q = ( pi , ai ), i = 1,..., k is the vector of unknown parameters. ( ) By derivation of the function L x q in order to each of these parameters and equating to zero each of the expressions obtained, we have the so-called likelihood equations: ¶L(x q ) =0 (8) ¶q The equations obtained are sometimes impossible to solve analytically or very difficult to resolve, so often resort to the use of iterative methods. However, not always the most common processes are able to respond to the scale of problems. In order to overcome these difficulties there is the EM algorithm of (Dempster, et al., 1977), which is the most widely used in solving equations that describe maximum likelihood. In particular if we have a sample size n and a mixture of two univariate normal components, the log-likelihood function is given by: n é - (x j -m1 ) 2 - ( x j - m2 ) ù 2 L(x q ) = å ln p 1 1 ê 2s12 + (1 - p ) e 2s 2 ú 2 e (9) j =1 ê 2p s 1 2p s 2 ú ë û 4 Practical method for detection finite mixture of normal distributions The method aims on the one hand prove the existence of mixtures of normal distributions and on the other hand achieving the adjustment of an appropriate 774 model. This approach will first check the detailed curves detected in the histogram of simple frequencies, and then estimate the parameters of these curves and the model checking by applying the Kolmogorov-Smirnov test, adapted to a mixture of normal. 4.1 Dissection of the Curves Consider the original data set x1 ,..., xn of the sample under study, and begin a first step by ordering them in ascending order. We then have the ordered sample z1 ,..., zn with z1 £ ... £ zn . Let n1 , n2 and n3 represent the sizes of sub- populations derived from the decomposition of the initial population, with n1 + n2 + n3 = n . Considering the hypothesis of existence of two normal distributions, denote 2n1 2n3 p= and 1 - p = where p and 1 - p represent the proportions of n n each one of the curves found on the original curve. These curves are normal, so they are symmetrical and therefore there is n1 + n2 1 = , or is 50% the size of the sample size n . Let m1 and s 1 be n 2 respectively the mean and standard deviation in the distribution curve 1 and m2 and s 2 be respectively the mean and standard deviation in the distribution curve 2. Thus the probability density function f of weighing the two normal curves is defined by f (x p , m1 , m2 ,s 1 ,s 2 ) = pn(x m1 ,s 1 ) + (1 - p )n(x m2 ,s 2 ) (10) 4.2 Estimation of parameters for the general model 4.2.1 Estimates of n1* The value of n1* is the one that minimizes D(n' ) with D(n' ) = x - ( p m '* n * 1,n' ( + 1- p m '* n ) ) , where x is the sample mean and * 2 ,n' 775 ì * z n' + z n' +1 ïm1,n' = 2 ï ï z n +z n ï * n' + n' + +1 ím 2 ,n' = 2 2 (11) ï 2 ï '* 2 n' ï pn = n ï î For each n' there is an estimate of the fraction correspondent to each sub- population pn and 1 - pn and of the respective mean values m1,n' and m 2 ,n' . '* '* * * Combining these estimates we obtain: mn'* = p'*n m1*,n' + (1 - p'*n )m2* ,n' (12) and the minimization of: D(n' ) = x - mn'* (13) 4.2.2 Estimates of p , m1 and m 2 Once estimated n1 we use the correspondent estimates for the fractions and the mean values of sub-populations obtaining: ì * z n1 + z n1 +1 ïm1 = ï 2 ï z n +z n ï * n1 + n1+ +1 ím 2 = 2 2 (14) ï 2 ï * 2n1 ïp = n ï î the distribution mean value will be given by m = pm1 + (1 - p )m2 . 4.2.3 Estimates of s 1 and s 2 776 To obtain the estimates for s 1 and s 2 we use: ì ï n ïn3 = 2 - n1 * ï * ï 2* 1 n1 ís 1 = * å z j - m1 n1 j =1 ( * 2 ) (15) ï ï ( ) n 1 ïs 22* = * å z j - m 2* 2 ï n3 j =n* + n +1 î 1 2 n * As we have seen the reason for equality n3* = - n1 and is justified by the fact 2 that the tails of two distributions assume negligible values. So, since n2 = n1 + n3 , and n2 = n 2 n ( it finally comes n3 = - n1 . The term z j - m1 reflects the 2 * ) differences between each value of sample 1. s 12* and s 22* denote the estimated variances for populations 1 and 2. 4.3 Validation of the Model We will use the Kolmogorov-Smirnov test to check the model. The statistic of this test is the maximum module of the difference between the empirical distribution ì0 ; x < z1 ïi ï Fx* í ; zi £ x £ zi +1 , i = 1,..., n (16) ïn ïî1 ; x > zn and the adjusted ( ) ( ) ( p* N x m1* ,s 1* + 1 - p* N x m2* ,s 2* . ) One of the observed values is then equal to ìi ( ( ) ( ) ( ü )) Maxí - p* N zi m1* ,s 1* + 1 - p* N x m 2* ,s 2* ý ; i = 1,..., n în þ 777 We will obtain the values of distributions ( N z ml* ,s l* ) ; l = 1,2 , in z i points, i = 1,..., n . From the zi points, i = 1,..., n , we calculate ì zi - m1* u ï i = ï s 1* í ; i = 1,..., n (17) ïv = zi - m 2 * ïî i s 2* Now since ( ) ì N zi m1* ,s 1* = N (ui 0,1) ; i = 1,..., n ï í ï ( * * ) î N zi m 2 ,s 2 = N (vi 0,1) ; i = 1,..., n The problem is reduced to obtain the values of standardized normal at ui , i = 1,..., n and at ui , i = 1,..., n . We have y2 - 2 j +1 1 ¥ (- 1) w w j N (w 0,1) = + 2 1 w e 1 w ò 2 w 0 2p dw = + 2 w å j 2p j =0 2 2 j + 1 (18) So, the problem is simplified since in general it is enough to handle the first 25 items. The above expression is justified since if w < 0 we have y2 y2 y2 w - 0 - w - 2 2 2 e 1 e 1 e N (w ) = ò dw = -ò dw = - ò dw (19) -¥ 2p 2 w 2p 2 0 2p and if w>0 y2 y2 y2 w - w - w - 2 2 2 e 1 e 1 e N (w ) = ò dw = +ò dw = + ò dw (20) 0 2p 2 0 2p 2 0 2p After obtaining the values for normal distribution we calculate the values: ( F* ( zj ) = p* N( uj 0 1, ) + 1 - p* N(v j 0,1) ) (21) 778 And we obtain the statistic ìj ü T = maxí - F * (z j ) ; j = 1,..., ný în þ This value should be compared with the value in normal distribution table and we concluded about the existence of a mixture of normal distributions if the T value is less than the tabulated. For the previous iterative procedure we decided to simplify the calculations, very time consuming, by designing an algorithm, and we used it in our practical application. 5 Application First, we will carry out a comprehensive statistical study of available data relating to populations of vine varieties. We will discuss three types of caste, namely the Trincadeira Preta in years 1988, 1989 and 1990 in the region of Almeirim, Aragon 1987, 1988 and 1989 in the region of Reguengos and Touriga Nacional 1994, 1995 and 1996 in the region of Foz Coa. Regarding the number of observations and replicates of clones for analysis we noted that with regard to the caste Trincadeira Preta there is in each year a total of (271 observations x 5 repetitions) , in Touriga Nacional (197 observations x 5 repetitions) and in the Aragonês (153 observations x 4 repetitions). Note that each observation is respective to a different clone, repeating this clone 4 or 5 times. For a better layout follows the table 1: Table 1. Varieties per year and repetitions Varieties Year Repetition Trincadeira Preta 1988,1989,1990 REP1;REP2;REP3;REP4;REP5 Aragonês 1987,1988,1989 REP1;REP2;REP3;REP4 Touriga Nacional 1994,1995,1996 REP1;REP2;REP3;REP4;REP5 For each repetition, we proceeded to calculate the mean, variance, standard deviation, median, sum of sample values, 95%, first quartile, third quartile, range of the sample, inter-quartile range, skewness coefficient, coefficient of flattening and determination of maximum and minimum values of the sample. Then we obtained the histograms for each of the repetitions, and we tested the normality through the Kolmogorov-Smirnov test. 5.1 Analysis of distributions of repeat genotypes 5.1.1 Testing the normality of distributions As the normal distribution in one of the most important ones, it is useful at this point proceed to test data normality. To this end we then base our conclusions on the results of a nonparametric test, as mentioned above, the application of the Kolmogorov-Smirnov test. 779 With the help of statistical software and consulting the Table of Critical Values of the Kolmogorov-Smirnov test for one sample, then we obtained the values KS Observed and KS. Tabulated for each repetition and present in table: Table 2. Kolmogorov-Smirnov application K.S. K.S. Tabulated Varietie Year Repetition Observed 1% 5% REP1 0.06582 <0.09902 <0.08261 REP2 0.08012 <0.09902 <0.08261 1988 REP3 0.06842 <0.09902 <0.08261 REP4 0.07813 <0.09902 <0.08261 REP5 0.08156 <0.09902 <0.08261 REP1 0.04241 <0.09902 <0.08261 Trincadeira REP2 0.04603 <0.09902 <0.08261 Preta 1989 REP3 0.05565 <0.09902 <0.08261 (n=271) REP4 0.06031 <0.09902 <0.08261 REP5 0.06810 <0.09902 <0.08261 REP1 0.05569 <0.09902 <0.08261 REP2 0.06024 <0.09902 <0.08261 1990 REP3 0.03547 <0.09902 <0.08261 REP4 0.06007 <0.09902 <0.08261 REP5 0.03681 <0.09902 <0.08261 Table 3. Kolmogorov-Smirnov application K.S. K.S. Tabulated Varietie Year Repetition Observed 1% 5% REP1 0.07008 <0.13178 <0.10995 REP2 0.07756 <0.13178 <0.10995 1987 REP3 0.07637 <0.13178 <0.10995 REP4 0.11603 <0.13178 >0.10995 REP1 0.08873 <0.13178 <0.10995 REP2 0.04576 <0.13178 <0.10995 Aragonês 1988 REP3 0.06797 <0.13178 <0.10995 (n=153) REP4 0.11283 <0.13178 >0.10995 REP1 0.09338 <0.13178 <0.10995 REP2 0.10146 <0.13178 <0.10995 1989 REP3 0.13492 >0.13178 >0.10995 REP4 0.12791 <0.13178 >0.10995 Table 4. Kolmogorov-Smirnov application K.S. K.S. Tabulated Varietie Year Repetition Observed 1% 5% REP1 0.07714 <0.11350 <0.09690 REP2 0.09881 <0.11350 >0.09690 1994 REP3 0.05603 <0.11350 <0.09690 REP4 0.05166 <0.11350 <0.09690 REP5 0.06040 <0.11350 <0.09690 REP1 0.06864 <0.11350 <0.09690 Touriga REP2 0.07562 <0.11350 <0.09690 Nacional 1995 REP3 0.06499 <0.11350 <0.09690 («=197) REP4 0.06426 <0.11350 <0.09690 REP5 0.08181 <0.11350 <0.09690 REP1 0.06146 <0.11350 <0.09690 REP2 0.08055 <0.11350 <0.09690 1996 REP3 0.08516 <0.11350 <0.09690 REP4 0.03227 <0.11350 <0.09690 REP5 0.06266 <0.11350 <0.09690 We considered 1.63 for 1% and 1.36 for 5% significance level, respectively. 780 In the analysis of the results, we note that the null hypothesis that the sample comes from a normal population, is rejected in five cases for the significance level of 5% and only in one case to significance level of 1%, which is not altogether surprising since we are facing a total of 42 cases. If we consider a binomial distribution with n = 42 and x = 5 then it's expected that 5% of observations fall outside the standard of reference, in our case the normal distribution. 6 Considerations and remarks The vegetative reproduction seems to guarantee the homogeneity of genotypes. So in a given year and local the productions of the same genotype should be distributed normally. Not always this happens because when we studied the 42 repetitions of genotypes: Aragonês, Trincadeira Preta and Touriga Nacional four cases were found where the theoretical model did not fit significantly. In these four cases it was possible to fit the data a mixture of two normal distributions. We consider important in future work using the techniques of ANOVA to estimate the variance components internal to the genotypes and between genotypes. References 1. Bhattacharya, C.G. (1967) A simple method for resolution of a distribution into its Gaussian components. Biometrics 2, p.115-135. 2. Cassie, R.M. (1954) Some uses of Probability Paper for the Graphical Analysis of Polymodal Frequency Distributions. Austral J. Marine and Freshwater Res. 5, p.513-522. 3. Dempster, A.P, Laird, N.M. and Rubin, D.B. (1977) Maximum likelihood from incomplete data via the EM algorithm (with discussion). J.R. Statist. Soc. B 39, p.1-38. 4. Fowlkes, E.B. (1979) Some Methods for Studying the Mixture of two normal (lognormal) distributions. JASA 74, p.561-575. 5. Harding, J.P. (1948) The Use of Probability for the Graphical Analysis of Polymodal Frequency Distributions. J. Marine Biol. Assoc. U.K. 28, p.141-153. 6. Oliveira, A. (1999) Misturas de Normais: Uma Aplicação. Tese de Mestrado. FCT/UNL. 7. Pearson, K. (1894) Contributions to the Mathematical theory of evolution. Phil. Trans. A. 185, p.71-110. 8. Quandt R. E. & Ramsey, J. B. (1978) Estimating Mixtures of Normal Distributions and Switching Regressions (with discussion). JASA, 73, p.730- 752. 9. Teicher, H. (1963) Identifiability of Finite Mixtures. Ann. Math. Statist., 34, p. 1265-1269. 781