- Used Books
- Staff Picks
- Gifts & Gift Cards
- Sell Books
- Stores & Events
- Let's Talk Books
Special Offers see all
More at Powell's
Recently Viewed clear list
Ships in 1 to 3 days
available for shipping or prepaid pickup only
Available for In-store Pickup
in 7 to 12 days
Other titles in the Stochastic Modelling and Applied Probability series:
A Probabilistic Theory of Pattern Recognitionby Luc Devroye
Synopses & Reviews
Pattern recognition presents one of the most significant challenges for scientists and engineers, and many different approaches have been proposed. The aim of this book is to provide a self-contained account of probabilistic analysis of these approaches. The book includes a discussion of distance measures, nonparametric methods based on kernels or nearest neighbors, Vapnik-Chervonenkis theory, epsilon entropy, parametric classification, error estimation, free classifiers, and neural networks. Wherever possible, distribution-free properties and inequalities are derived. A substantial portion of the results or the analysis is new. Over 430 problems and exercises complement the material.
A self-contained and coherent account of probabilistic techniques, covering: distance measures, kernel rules, nearest neighbour rules, Vapnik-Chervonenkis theory, parametric classification, and feature extraction. Each chapter concludes with problems and exercises to further the readers understanding. Both research workers and graduate students will benefit from this wide-ranging and up-to-date account of a fast- moving field.
Includes bibliographical references (p. -618) and indexes.
Table of Contents
Preface * Introduction * The Bayes Error * Inequalities and alternate distance measures * Linear discrimination * Nearest neighbor rules * Consistency * Slow rates of convergence Error estimation * The regular histogram rule * Kernel rules Consistency of the k-nearest neighbor rule * Vapnik-Chervonenkis theory * Combinatorial aspects of Vapnik-Chervonenkis theory * Lower bounds for empirical classifier selection * The maximum likelihood principle * Parametric classification * Generalized linear discrimination * Complexity regularization * Condensed and edited nearest neighbor rules * Tree classifiers * Data-dependent partitioning * Splitting the data * The resubstitution estimate * Deleted estimates of the error probability * Automatic kernel rules * Automatic nearest neighbor rules * Hypercubes and discrete spaces * Epsilon entropy and totally bounded sets * Uniform laws of large numbers * Neural networks * Other error estimates * Feature extraction * Appendix * Notation * References * Index
What Our Readers Are Saying
Other books you might like
Reference » Science Reference » General