14.2. Mixture Decomposition Schemes

The basic reasoning behind this algorithmic family springs from our familiar Bayesian philosophy. We assume that there are m clusters, Cj, j = 1, …, m, underlying the data set.[2] Each vector xi, i = 1, …, N, belongs to a cluster Cj with probability P(Cj|xi). A vector xi is appointed to the cluster Cj if

2 Recall that the number m is assumed to be known.

The differences from the classification task of Chapter 2 are that (a) no training data with known cluster labeling are available and (b) the a priori probabilities P(Cj)Pj are not known either. Thus, although the goal is the same, the tools have to be ...

Get Pattern Recognition, 4th Edition now with the O’Reilly learning platform.

O’Reilly members experience books, live events, courses curated by job role, and more from O’Reilly and nearly 200 top publishers.