- Hardcover: 680 pages
- Publisher: Wiley-Interscience; 2 edition (November 9, 2000)
- Language: English
- ISBN-10: 0471056693
- ISBN-13: 978-0471056690
- Product Dimensions: 7.1 x 1.3 x 10.1 inches
- Shipping Weight: 2.9 pounds (View shipping rates and policies)
- Average Customer Review: 46 customer reviews
- Amazon Best Sellers Rank: #170,295 in Books (See Top 100 in Books)
Enter your mobile number or email address below and we'll send you a link to download the free Kindle App. Then you can start reading Kindle books on your smartphone, tablet, or computer - no Kindle device required.
To get the free app, enter your mobile phone number.
Pattern Classification (Pt.1) 2nd Edition
Use the Amazon App to scan ISBNs and compare prices.
The Amazon Book Review
Author interviews, book reviews, editors picks, and more. Read it now
Frequently bought together
Customers who bought this item also bought
"…it provides a good introduction to the subject of Pattern Classification." (Journal of Classification, September 2007)
"…a fantastic book! The presentation...could not be better, and I recommend that future authors consider…this book as a role model." (Journal of Statistical Computation and Simulation, March 2006)
"...strongly recommended both as a professional reference and as a text for students..." (Technometrics, February 2002)
"...provides information needed to choose the most appropriate of the many available technique for a given class of problems." (SciTech Book News, Vol. 25, No. 2, June 2001)
"I do not believe anybody wishing to teach or do serious work on Pattern Recognition can ignore this book, as it is the sort of book one wishes to find the time to read from cover to cover!" (Pattern Analysis & Applications Journal, 2001)
"This book is the unique text/professional reference for any serious student or worker in the field of pattern recognition." (Mathematical Reviews, Issue 2001k)
"...gives a systematic overview about the major topics in pattern recognition, based whenever possible on fundamental principles." (Zentralblatt MATH, Vol. 968, 2001/18)
"attractively presented and readable" (Journal of Classification, Vol.18, No.2 2001)
From the Back Cover
From the reviews . . .
"The first edition of this book, published 30 years ago by Duda and Hart, has been a defining book for the field of Pattern Recognition. Stork has done a superb job of updating the book. He has undertaken a monumental task of sifting through 30 years of material in a rapidly growing field and presented another snapshot of the field, determining what will be of importance for the next 30 years and incorporating it into this second edition. The style is easy to read as in the original book and the statistical, mathematical material comes alive with many new illustrations. The end result is harmonious, leading the reader through many new topics..."
–Sargur N. Srihari, PhD, Director, Center for Excellence in Document Analysis and Recognition, Distinguished Professor, Department of Computer Science and Engineering, SUNY at Buffalo
Practitioners developing or investigating pattern recognition systems in such diverse application areas as speech recognition, optical character recognition, image processing, or signal analysis, often face the difficult task of having to decide among a bewildering array of available techniques. This unique text/professional reference provides the information you need to choose the most appropriate method for a given class of problems, presenting an in-depth, systematic account of the major topics in pattern recognition today. A new edition of a classic work that helped define the field for over a quarter century, this practical book updates and expands the original work, focusing on pattern classification and the immense progress it has experienced in recent years. Special features include:
- Clear explanations of both classical and new methods, including neural networks, stochastic methods, genetic algorithms, and theory of learning
- Over 350 high-quality, two-color illustrations highlighting various concepts
- Numerous worked examples
- Pseudocode for pattern recognition algorithms
- Expanded problems, keyed specifically to the text
- Complete exercises, linked to the text
- Algorithms to explain specific pattern-recognition and learning techniques
- Historical remarks and important references at the end of chapters
- Appendices covering the necessary mathematical background
NOTE: Computer Manual in MATLAB to Accompany Pattern Classification, 2e users access toolbox via ftp://ftp.wiley.com/public/sci_tech_med/pattern_classification/ (Note: Visitors will require a password from the Manual to access.)
Discover books for all types of engineers, auto enthusiasts, and much more. Learn more
Top customer reviews
With this in mind the authors and their new coauthor David Stork go about the task of providing a revision. True to the goals of the original the authors undertake to describe pattern recognition under a variety of topics and with several available methods to cover each topic. Important new areas are covered and old but now deemed less significant are dropped. Advances in statistical computing and computing in general also dictate the topics. So although the authors are the same and the title is almost the same (note that scene analysis is dropped from the title) it is more like an entirely new book on the subject rthan a revision of the old. For a revision, I would expect to see mostly the same chapters with the same titles and only a few new chapters along with expansion of old chapters.
Although I view this as a new book, that is not necessarily bad. In fact it may be viewed as a strength of the book. It maintains the style and clarity of the original that we all loved but represents the state-of-the-art in pattern recognition at the beginning of the 21st Century.
The original had some very nice pictures. I liked some of them so much that I used them with permission in the section on classification error rate estimation in my bootstrap book. This edition goes much further with beautiful graphics including many nice three-dimensional color pictures like the one on the cover page.
The standard classical material is covered in the first five chapters with new material included (e.g. the EM algorithm and hidden markov models in Chapter 3). Chapter 6 covers multilayer neural networks (a totally new area). Nonmetric methods including decision trees and the CART methodology are covered in Chapter 8. Each chapter has a large number of relevant references and many homework exercises and computer exercises.
Chapter 9 is "Algorithm-Independent Machine Learning" and it includes the wonderful "No Free Lunch" theorem (Theorem 9.1), a discussion of the minimum desciption length principle, overfitting issues and Occam's razor, bias - variance tradeoffs,resampling method for estimation and classifier evaluation, and ideas about combining classifiers.
Chapter 10 is on unsurpervised learning and clustering. In addition to the traditional techniques covered in the first edition the authors include the many advances in mixture models.
I was particularly interested in that part of Chapter 9. There is good coverage of the topics and they provide a number of good references. However, I was a bit disappointed with the cursory treatment of bootstrap estimation of classification accuracy (section 9.6.3 on pages 485 - 486). I particularly disagree with the simplistic statement "In practice, the high computational complexity of bootstrap estimation of classifier accuracy is rarely worth possible improvements in that estimate (Section 9.5.1)". On the other hand, the book is one of the first to cover the newer and also promising resampling approaches called "Bagging" and "Boosting" that these authors seem to favor.
Davison and Hinkley's bootstrap text is mentioned for its practical applications and guidance for bootstrapping. The authors overlook Shao and Tu which offers more in the way of guidance. Also my book provides some guidance for error rate estimation but is overlooked.
My book also illustrate the limitations of the bootstrap. Phil Good's book provides guidance and is mentioned by the authors. But his book is very superficial and overgeneralized with respect to guiding practitioners. For these reasons I held back my enthusiasm and only gave this text four stars.
For a more up-to-date treatment, see McLachlan's recent book in the Wiley statistics series. However, this book provides valuable explanations of Bayes rules and shows pictorially what the boundaries look like for linear and quadratic classifiers. In fact I borrowed their pictures in Chapter 2 of my book on bootstrap methods and it appears in both the first and second editions of my book.
The authors with the help of a third author have updated this book recently and I highly recommend the new addition which maintains many of the nice features of the original.