2.2. Bayes Decision Theory

We will initially focus on the two-class case. Let ω1, ω2 be the two classes in which our patterns belong. In the sequel, we assume that the a priori probabilities P1), P2) are known. This is a very reasonable assumption, because even if they are not known, they can easily be estimated from the available training feature vectors. Indeed, if N is the total number of available training patterns, and N1, N2 of them belong to ω1 and ω2, respectively, then P1) ≈ N1/N and P2) ≈ N2/N.

The other statistical quantities assumed to be known are the class-conditional probability density functions p(xi), i = 1, 2, describing the distribution of the feature vectors in each of the classes. If these are not known, they ...

Get Pattern Recognition, 4th Edition now with the O’Reilly learning platform.

O’Reilly members experience books, live events, courses curated by job role, and more from O’Reilly and nearly 200 top publishers.