MML mixture modelling of rnulti-state, Poisson, von Mises circular and Gaussian distributions

Chris S. Wallace and David L. Dowe Department of Computer Science, Monash University, Clayton, Victoria 3168, Australia e-mail: {csw, dld} @cs.monash.edu.au

Abstract Bayes's Theorem. Since D and Pr(D) are given and we wish to infer H , we can regard the problem of maximis­ Minimum Message Length (MML) is an invariant ing the posterior probability, Pr(H!D), as one of choos­ Bayesian point estimation technique which is also con­ ing H so as to maximise Pr(H). Pr(DIH) . sistent and efficient. We provide a brief overview of An information-theoretic interpretation of MML is MML inductive inference (Wallace and Boulton (1968) , that elementary coding theory tells us that an event of Wallace and Freeman (1987)) , and how it has both probability p can be coded (e.g. by a Huffman code) by an information-theoretic and a Bayesian interpretation. a message of length l = - log2 p bits. (Negligible or no We then outline how MML is used for statistical pa­ harm is done by ignoring effects of rounding up to the rameter estimation, and how the MML mixture mod­ next positive integer.) elling program, Snob (Wallace and Boulton (1968), Wal­ So, since - log2 (Pr(H). Pr(D!H)) = - log2 (Pr(H)) - lace (1986) , Wallace and Dowe(1994)) uses the message log2 (Pr(DjH)), maximising the posterior probability, lengths from various parameter estimates to enable it to Pr(HID), is equivalent to minimising combine parameter estimation with selection of the num­ ber of components. The message length is (to within MessLen = -log2(Pr(H))-log2 (Pr(DIH)) (1) a constant) the logarithm of the posterior probability the length of a two-part message conveying the theory, of the theory. So, the MML theory can also be re­ H , and the data, D, in light of the theory, H . Hence garded as the theory with the highest posterior proba­ the name "minimum message length" (principle) for thus bility. Snob currently assumes that variables are uncor­ choosing a theory, H , to fit observed data, D. The prin­ related, and permits multi-variate data from Gaussian, ciple seems to have first been stated by Solomonoff[31, discrete multi-state, Poisson and von Mises circular dis­ p20] and was re-stated and apparently first applied in a tributions. series of papers by Wallace and Boulton[37, p185][4][5, pp63-64][6, 8, 7, 38] dealing with model selection and parameter estimation (for Normal and multi-state vari­ 1 Introduction - About Minimum ables) for problems of mixture modelling (also known as Message Length (MML) clustering, numerical taxonomy or, e.g. [3], "intrinsic classification"). The Minimum Message Length (MML)[37, pl85][43] An important special case of the Minimum Message (and, e.g., [5, pp63-64][38]) principle of inductive infer­ Length principle is an observation of Chaitin[9] that data ence is based on information theory, and hence lies on the can be regarded as "random" if there is no theory, H, de­ interface on computer science and statistics. A Bayesian scribing the data which results in a shorter total message interpretation of the MML principle is that it variously length than the null theory results in. For a comparison states that the best conclusion to draw from data is the with the related Minimum Description Length (MDL) theory with the highest posterior probability or, equiv­ work of Rissanen[28, 29], see, e.g., [32]. We discuss later alently, that theory which maximises the product of the some (other) applications of MML. prior probability of the theory with the probability of the data occuring in light of that theory. We quantify by this immediately below. 2 Parameter Estimation Letting D be the data and H be an hypothesis (or MML theory) with prior probability Pr(H), we can write the posterior probability Pr(H!D) = Pr(H&D)/Pr(D) Given data x and parameters B, let h(B) be the prior Pr(H). Pr(DjH)/ Pr(D), by repeated application of on B, let p(xlB) be the likelihood,

529 let L = - logp(xlB) be the negative log-likelihood and 2.1 Gaussian Variables let (2) For a No rm al distribution (with sample size, N), as­ suming a uniform prior on µ and a scale-invariant, be the Fisher information, the determinant of the (Fisher l/cr prior on er , we get that the Maximum Likelihood information) matrix of expected second partial deriva­ (ML) and MM L estimates of the mean concur, i.e., that 2 2 tives of the negative log-likelihood. Then the MML es­ fl.MML =µ ,'VIL= i:. Letting s = L ;(x; - x) , we get timate of B [43, p245] is that value of Bminimising the that cr 2 ML = s2 /N and[37, p190) that message length, •) 2

531 le. For each component, the distribution parameters assignment of things to components. The reason for this of the component (as discussed in Sect.ion 2) . Ea.ch pa­ is that(33, §3](34, p 77] if p(j, x), j = 1, . ., .J , is the proba­ rameter is considered to be specified to a precision of bility of component j generating datum x , then the total the order of its expected estimation error or uncertainty assignment of x to its best component results in a mes­ (see Section 2 or, e.g. , (39, pp3-4)). For a larger compo­ sage length of - log(maxj p(j, x)) to enco de x whereas, nent, the parameters will be encoded to greater precision letting P(x) = Lj p(j, x), a partial assignment of x hav­ and hence by longer fragments than for a less abundant ing probability p(j,x)/P(x) of being in component j re­ component. sults in a shorter message length of - log( P( x)) to en­ ld. For each thing, the component to which it is esti­ code x. As shown in [33 , §3](34, p77][41] , this shorter mated to belong. (This can be done using the Huffman length is achievable by a message which asserts definite code referred to in 1(b) above.) membership of each thing by use of a special coding trick. Having stated in part 1 of the message above, our hypothesis, H , about how many components there are and what the distribution parameters (µ, r:r , etc.) are 4 Consistency, invariance and ef­ for each attribute for each component, in part 2 of the ficiency of MML estimates message we need to state the data, D , in light of this ·hypothesised model, H . If the outcomes of any random process are encoded us­ The details of the encoding and of the calculation of ing a code that is optimal for that process, the result­ the length of part 1 of the message may be found in ing binary string forms a completely random process(43, Section 2 and elsewhere(37 , 39] . It is perhaps worth not­ p241]. This fact and the fact that general MML codes ing here that since our objective is to minimise the mes­ are (by definition) optimal implicitly suggest that, given sage length (and maximise the posterior probability), we sufficient data, MML will converge as closely as possible never need construct a message - we only need be able to any underlying model. Indeed , MML can be thought to calculate its length. of as extending Chaitin's idea of randomness[9] to al­ Given that part 1( d) of the message told us which com­ ways trying to fit given data with the shortest possible ponent each thing was estimated to belong to and that, computer program (plus noise) for generating it. This for each component, part l(c) gives us the (MML) esti­ general convergence result for MML has been explicitly mates of the distribution parameters for each attribute, re-stated elsewhere[36 , l]. Similar arguments show that part 2 of the message now encodes each attribute value MML estimates are not only consistent, but that they of each thing in turn in terms of the distribution param­ are also efficient, i.e., that they converge to any true un­ eters (for each attribute) for the thing's component. derlying parameter value as quickly as possible. The fact that VF transforms like a prior is a basis 3.2 Stating the message more concisely used by some to choose VF as a Jeffrey's prior. Al­ using partial assignment though we do not wish to advocate the use of a Jeffrey's prior, we do note that h/VF is invariant under parame­ Part 1( d) of the message described in the previous ter transformation. Since the likelihood function is also section (§3.l) implicitly restricts us to hypotheses, H , invariant under parameter transformation, we see from which assert with 100% definiteness which component equation (3) that MML is also invariant under parameter each thing belongs to. Given that the population that transformation[43, 38]. we might encounter could consist of two different but The problem of model selection and parameter estima­ highly over-lapping distributions, forcing us to state def­ tion in mixture modelling can, at its worst, be thought initely which component each thing belongs to is bound of as a problem for which the number of parameters to cause us to mis-classify outliers from one distribu­ to be estimated grows with the data. It is well known tion as belonging to another. In the case of two over­ that Maximum Likelihood can become inconsistent (or lapping (but distinguishable) 1-dimensional Normal dis­ very inefficient) with such problems, e.g. multiple factor tributions, this would cause us to over-estimate the dif­ analysis[35] and the Neyman-Scott problem[24, 16]. ference in the component means and under-estimate the component standard deviations. Since what we seek is a message which enables us to 5 Alternative Bayesian methods encode the attribute values of each thing as concisely as possible, we note that a shorter message than that of In doing inductive inference of mixture models from Section 3.1 can be obtained by a probabilistic (or partial) data, there are several levels of inference that we might conceivably wish to make. We might wish simply to in­ Furthermore, whereas MML is known to be fer the most likely number of components. Or, alterna­ invariant[:38, 43) under l-to-1 transformations, the MAP tively, we might wish to infer the number of components, (posterior ) estimate is known generally not to be their relative abundances and the parameter values as­ invariant under 1-to-1 transformations - e.g. , von Mises sociated with each component. Or, we might further circular parameter estimation[l4) in polar and Cartesian wish to infer the above and a probabilistic assignment co-ordinates. of things to components. It is these last two variations While the authors do not advocate MAP, another which are variously understood by the term "mixture Bayesian method which the authors do advocate Is es­ modelling". Finally, one might wish to infer the number timation by minimising the Expected Kullback-Leibler of components and the identities of their members with­ distance (min EKL) . Like the MML estimator, min EKL out regard to parameter estimation. This form is often is invariant under re-parameterisation. Work to appear5 termed "clustering". follows Wallace[36] and shows strong similarities between MAP (maximum a posteriori) operates on a density Strict MML[38, 43] and min EKL (as is easily seen in the and must marginalise over (or integrate out) parameters case of M-state Bernoulli sampling). to estimate memberships, and must likewise marginalise over memberships to estimate parameters. MAP (like penalised likelihood methods) is unable consistently to 6 Alternative mixture modelling estimate both parameter values and class memberships. programs Let us see why this is: consider some estimate of the number of components followed by parameter estimates The first Snob program (since out-dated)(37) was possi­ for each of these components. (We could, for example, bly the first program for Gaussian mixture modelling, have two equally abundant and substantially overlapping although many statistical and machine learning ap­ I-dimensional Normal distributions with the same stan­ proaches to this problem have been developed since (e.g., dard deviation, a- .) If we assign each thing to its most McLachlan et al.(23, 22), D. Fisher's CobWeb(17)). Dis­ probable class, there will be a neat division of things to cussions of early alternative algorithms for Gaussian classes, a division which will not be consistent with the mixture modelling have been given by Boulton(3). original estimates of means and a- . Rather than obtain probabilities from densities of 6.1 Comparison with AutoClass II real-valued parameters by integrating (as MAP does) , MML obtains such probabilities by rounding-off (or Like Snob, AutoClass II [10] assumes6 a prior distribu­ quantising) 2 the possible parameter estimates into cod­ tion over the number of classes and independent prior ing blocks (or uncertainty regions) as discussed in Sec­ densities over the distribution parameters of the sample tion 2. By shortening the length of the message to a class densities. However[34), AutoClass II is not based minimum, MML arrives at the (quantised) theory of the on a message length criterion, but instead makes a more highest probability (see Section 1) whose resulting binary direct inference of the number of classes, J . string forms[43, p 241][41, p 41] a completely random Let V be the vector of abundance and distribution process. The fact that the first part of the message parameters needed to specify a model with J compo­ string3 and the second part of the message are com­ nents. Let P(J) be the prior probability of having J pletely random (and "noise" ) means that the coding components, and let h(V) be the prior probability of the trick4 causes the assignment of data things to compo­ parameters, V. Let X denote the data, i.e. the set of nents to be done (pseudo-)randomly in a way which is attribute values for all things, and let P(X!V) be the consistent with the parameter estimates. If we do not probability of obtaining data X given the ]-component minimise the message length (by taking advantage of the model specified by V. The joint probability P(J, X) of coding trick), as with MAP estimation, inconsistencies J and X is then will arise. P(J, X) = h(V)P(XjV)dV (9) Results of Barron and Cover[l] show MML to be con­ J sistent for any i.i.d. problem, and other results[36][43, p and the posterior probability, P(JIX), of J given the 241] show MML (and Strict MML[38, 43]) to be consis­ data, X , is tent and efficient for problems of arbitrary generality. P(JIX) = P(J, X)/C£ P(j, X)) (10) 2 hence, Peter Cheeseman (private communication) refers to j MML as "quantised Bayes" 3 and part Id in particular see Section 3.1 5 by the current authors, and R. Baxter and J . Oliver 4 see Section 3.2 6 This sub-section is very much a re-writing of (34, pp78-80].

533 The calculation of the posterior, P(JIX), requires protein dihedral angles[12] . The Poisson module is cur­ t.he calculation of an integral for each possible num­ rently being used to model run lengths of helices and ber of classes, .] , in order to obtain the joint proba­ other protein conformations as being a mixture of Pois­ bility, P(.J, X). The integrand is proportional to the son distributions. This work should indirectly lead to a posterior density of the parameters of a .I-class model, better way of predicting protein conformations. h.(V) x P(XjV). Extensive surveys of Snob applications are given in Au toClass II approximates the integral by making the Patrick(27] and Wallace and Dowe[41] , with a recent ap­ assumption that most of the contribution to the integral plication of Gaussian mixture modelling to data on mem­ will come from the neighbourhood of the highest peak bers of grieving families is given in Kissane et al.[20). value of the integrand. It effectively fits a Gaussian func­ In applying Snob, a difference of more than 5 to tion to the integrand at this peak and uses the integral 6 bits[43, p25 l] or of more than 10 bits[33] might be of the Gaussian as its estimate of the true integral. Let­ deemed to be statistically significant under certain mod­ ting F be the Fisher information (from Section 2), the elling conditions. estimate is very similar, both analytically and numeri­ As well as having been applied to mixture models (dis­ cally, to the quantity h(V) x P(XjV)/VF, which is what cussed here), MML has also been successfully applied to MML (in general) and Snob (in particular) endeavour to a variety of problems of parameter estimation[37, 38, 43, maximise. Thus, although AutoClass II is differently 36, 39 , 40 , 14, 16], hypothesis testing(43, 39], Hidden motivated from Snob, in practice it gives almost identi­ Markov Models[19] and other multi-variate models[43, cal results. 36, 44 , 35, 16] . Further references are given in [13].

6.2 Comparison with other methods 8 Notes on further work and Oliver et al.[25) re-wrote the Gaussian mixture modelling Snob program extensions part of Snob[41 , 42) by modifying the Bayesian priors and introducing lattice constants[43, 39] (see Section 2.5) The Snob program currently implicitly assumes that and then empirically showed a successful performance variables are independent and uncorrelated. This could of (this slightly modified) Snob against AIC (Akaike's be modified to permit single linear (Gaussian) fac­ Information Criterion), BIC [28) and other methods. tor analysis[44) or multiple linear (Gaussian) factor The literature does not yet seem to contain any al­ analysis[35], or to model correlations via an inverse ternative algorithms for mixture modelling of von Mises Wishart or some other such prior. circular and Poisson distributions. It would not be too difficult[41) to permit the user to In general, with problems such as mixture modelling or modify the colourless priors (see Section 2) used by Snob multiple factor analysis where the number of parameters to better represent the user's prior beliefs (or knowled·ge , to be estimated increases with (and is potentially pro­ or bias). portional to) the amount of data, one must beware Max­ MML estimators have been obtained for the spheri­ imum Likelihood and MAP methods, which are both cal Fisher distribution[15] and work is currently under­ liable[24, 16] to give inconsistent results. way[26) to deal with the mixture modelling of these. When there are two or more overlapping components, a slight inefficiency will arise in the message length cal­ 7 Snob (and MML) Applications culations since parameters will be stated to a slightly higher than necessary degree of precision. The correc­ Earlier applications of Snob include several to medi­ tion for this can be computationally very slow and has cal, psychological, biological and exploratory geological been inspected in the Gaussian case by Baxter[2]. data, with a survey in (41] . The Poisson module seems to be accurately able to discriminate between pseudo­ randomly generated classes from different Poisson dis­ 9 Availability of the Snob pro­ tributions. It has also been used to analyse word-counts gram from a data-set of 17th Century texts. On this data­ set, a shorter message length was obtained by using a The current version of the Snob program (written Normal model than a Poisson model, and hence MML in Fortran 77) is freely available for not-for-profit, advocated the Normal model. The von Mises module academic research, and not for re-distribution, from has found clusters in data of several thousand sets of ftp://ftp.cs.monash.edu.au/pub/snob/Snob.README (or fr om c; _S . Wallace)_ Published or otherwise recorded (11] J .H. Conway and N ..J .A Sloane. Sphere Pack­ work using Snob should cite the current paper. User ings, Lattices and Groups. Springer-Verlag, London , guidelines are given in[41] and in the documentation file, 1988. snob.doc . (12] D.L. Dowe , L Allison, T.I. Dix, L. Hunter, C.S . Wallace, and T. Edgoose. Circular clustering of pro­ 10 Acknowledgments Section tein dihedral angles by Minimum Message Length. In Proceedings of the 1st Pacific Symposium on Bio­ This work was supported by Australian Research Coun­ co mputing (PSB-1), pages 242-255, Hawaii, U.S.A., cil (A RC) Grant A49330656 and ARC 1992 small grant January 1996. 9103169 . The authors wish to thank Graham Farr for valuable feedback. (13] D.L. Dowe and K.B. Korb. Conceptual difficulties with the Efficient. Market Hypothesis: Towards a naturalized economics. In D.L. Dowe, K.B. Korb, References and J ..J. Oliver. editors, Proceedings of the Infor­ mation, Statistics and Induction in Science (ISIS) (1] A.R. Barron and T.M. Cover. Minimum complexity Conference, pages 212-223, Melbourne, Australia, density estimation. IEEE Transactions on Informa­ August 1996. World Scientific. tion Theo ry, 37: 1034-1054, 1991. [14] D.L. Dowe. J ..J. Oliver, R .A. Baxter, and C.S. Wal­ (2] R.A. Baxter. Finding Overlapping Distributions lace. Bayesian estimation of the von Mises .concen­ with MML . Technical Report 95/244, Dept. of tration parameter. In Proc. 15th Maximum Entropy Computer Science, Monash University, Clayton Conference, Santa Fe , New Mexico, August 1995. 3168, Australia, November 1995. [15] D.L. Dowe, J ..J. Oliver, and C.S. Wallace. MML es­ (3] D.M. Boulton. The Information Criterion for In­ timation of the parameters of the spherical Fisher trinsic Classification. PhD thesis, Dept. Computer distribution. In A. et al. Sharma, editor, Proc. Science, Monash University, Australia, 1975. 7th Conj. Algorithmic Learning Theory (ALT'96), LNAI 1160, pages 213-227, Sydney, Australia, Oc­ [4] D.M. Boulton and C.S. Wallace. The information tober 1996. content of a multistate distribution. Journal of The­ oretical Biology, 23 :269-278, 1969. [16] D.L. Dowe and C.S. Wallace. Resolving the Neyman-Scott problem by Minimum Message (5] D.M. Boulton and C.S. Wallace. A program for Length. In Proc. Sydney International StatistiCal numerical classification. Computer Journal, 13:63- Congess (SISC-96) , pages 197-198, Sydney, Aus­ 69, 1970. tralia, 1996. [6] D.M. Boulton and C.S. Wallace. An information [17] D.H. Fisher. Conceptual clustering, learning from measure for hierarchic classification. The Computer examples, and inference. In Ma chine Learning: Journal, 16:254-261, 1973. Proceedings of the Fourth International Workshop, [7] D.M . Boulton and C.S. Wallace. An information pages 38-49. Morgan Kaufmann, 1987. measure for single-link classification. The Computer [18] N .I. Fisher. Statistical Analysis of Circular Data. Journal, 18(3):236-238, August 1975. Cambridge University Press, 1993 . . [8] D.M. Boulton and C.S. Wallace. A comparison be­ tween information measure classification. In Pro­ [19] M.P. Georgeff an.cl C.S. Wallace. A general crite­ ceedings of ANZAAS Con_gress , Perth, August 1973. rion for inductive inference. In T. O'Shea, editor, Advances in Artificial Intelligence : Proc. Sixth Eu­ (9] G .J . Chai tin. On the length of programs for com­ ropean Conference on Artificial Intelligence, pages puting finite sequences. Journal of the Association 473-482, Amsterdam, 1984. North Holland. for Computing Machinery, 13:547-549, 1966. [20] D. W . Kissane, S. Bloch, D. L. Dowe, R. D. Snyder, (10] P. Cheeseman, M. Self, J. Kelly, W . Taylor, D. Free­ P. Onghena, D. P. McKenzie, and C. S. Wallace. man, and J. Stutz. Bayesian classification. In Sev­ The Melbourne family grief study, I: Perceptions of enth National Conference on Artificial Intelligence, family functioning in bereavement. American Jour­ pages 607-611 , Saint Paul, Minnesota, 1988. nal of Psychiatry, 153:650-658, 1996.

535 (21] K.V . Mardia. Statistics of Directional Data. Aca­ - !CCI '90, pages 72 --· 81. Springer-Verlag, Berlin , demic Press, 1972. 1990.

[22] G.J. McLachlan Discriminant Analysis and Statis­ [35] C.S. Wallace. Multiple Factor Analysis by MML Es­ tical Pattern Recognition. Wiley, New York, 1992. timation. Technical Report 9.5/218, Dept. of Com­ puter Science, Monash University, Clayton, Victo­ (23] G.J. McLachlan and K.E. Basford. Mixture Models. ria 3168, Australia, 1995. submitted to .J . Multiv. Marcel Dekker, New York , 1988. Analysis. [24] J . Neyman and E.L. Scott. Consistent esti- [36] C.S. Wallace. False Oracles and SMML Estima­ mates based on partially consistent observations. tors. In D.L. Dowe , K.B . Korb , and .J.J . Oliver, Econometrika, 16:1 --32, 1948. editors, Proceedings of the Information, Statistics and Induction in Science (ISIS) Conference, pages (25] J. Oliver, Baxter R., and Wallace C. Unsupervised 304-316, Melbourne, Australia, August 1996. World learning using MML. In Proc. 13th International Scientific. Was Tech Rept 89/128, Dept. Comp. Sci ., Conj. Machine Learning (ICML 96), pages 364-372. Monash Univ., Australia . .June 1989. Morgan Kaufmann, San Francisco, CA , 1996. (37] C.S. Wallace and D.M . Boulton. An informa­ (26] J.J. Oliver and D.L. Dowe. Minimum Message tion measure for classification. Computer Journal, Length Mixture Modelling of Spherical von Mises­ 11:185-194, 1968 . Fisher distributions. In Proc. Sydney International Statistical Congess (SISC-96), page 198, Sydney, [38] C.S. Wallace and D.M. Boulton. An invariant Bayes Australia, 1996. method for point estimation. Classification Society Bulletin, 3(3):11-34, 1975. · [27] J .D. Patrick. Snob: A program for discriminat­ ing between classes. Technical Report TR 91/151, [39] C.S. Wallace and D.L. Dowe. MML estimation of Dept. of Computer Science, Monash University, the von Mises concentration parameter. Techni­ Clayton, Victoria 3168, Australia, 1991. cal report TR 93/193, Dept. of Computer Science, Monash University, Clayton, Victoria 3168, Aus­ [28] J. J. Rissanen. Modeling by shortest data descrip­ tralia, 1993. prov. accepted, Aust. J . Stat. tion. Automatica, 14:465-471, 1978. (40] C.S. Wallace and D.L. Dowe. Estimation of the (29] J . J. Rissanen. Stochastic Complexity in Statistical von Mises concentration parameter using Minimum Inquiry. World Scientific, Singapore, 1989. Message Length. In Proc. 12th Australian Statistical Soc. Conf., Monash University, Australia, 1994. (30] G. Schou. Estimation of the concentration param­ eter in von Mises-Fisher distributions. Biometrika, (41] C.S. Wallace and D.L. Dowe. Intrinsic classifica­ 65:369-377, 1978. tion by MML - the Snob program. In C. Zhang, J. Debenham, and D Lukose, editors, Proc. 7th [31] R.J. Solomonoff. A formal theory of inductive in­ Australian Joint Conj. on Artificial Intelligence, ference. Information and Control, 7:1-22,224-254, pages 37.:..44_ World Scientific, Singapore, 1994. See 1964. ftp://ftp.cs._monash.edu.au/pub/snob/Snob.README. (32] R.J. Solomonoff. The discovery of algorithmic prob­ [42] C.S. Wallace and D.L. Dowe. MML mixture mod­ ability: A guide for the programming of true cre­ elling of Multi-state, Poisson, von Mises circular ativity. In P. Vitanyi, editor, Computational Learn­ and Gaussian distributions. In Proc. Sydney Inter­ ing Theory: EuroCOLT'95, pages 1-22. Springer­ national Statistical Congess (SISC-96), page 197, Verlag, 1995. Sydney, Australia, 1996. _[33] C.S. Wallace. An improved program for classifica­ (43] C.S. Wallace and P.R. Freeman. Estimation and tion. In Proceedings of the Nineteenth Australian inference by compact coding. Journal of the Royal Computer Science Conference (ACSC-9), volume 8, Statistical Society (Series B), 49:240-252, 1987. pages 357-366, Monash University, Australia, 1986. (44] C.S. Wallace and P.R. Freeman. Single factor anal­ [34] C.S. Wallace. Classification by Minimum-Message­ ysis by MML estimation. Journal of the Royal Sta­ Length inference. In G. Goos and J. Hartmanis, tistical Society (Series B), 54:195-209, 1992. editors, Advances in Computing and Information