The book provides an overview of recent developments in large margin classifiers, examines connections with other methods (e.g., Bayesian inference), and identifies strengths and weaknesses of the method, as well as directions for future research. The concept of large margins is a unifying principle for the analysis of many different approaches to the classification of data from examples, including boosting, mathematical programming, neural networks, and support vector machines. The fact that it is the margin, or confidence level, of a classification-that is, a scale parameter-rather than a raw training error that matters has become a key tool for dealing with classifiers. This book shows how this idea applies to both the theoretical analysis and the design of algorithms. The book provides an overview of recent developments in large margin classifiers, examines connections with other methods (e.g., Bayesian inference), and identifies strengths and weaknesses of the method, as well as directions for future research. Among the contributors are Manfred Opper, Vladimir Vapnik, and Grace Wahba.
Publisher: MIT Press Ltd
Number of pages: 422
Weight: 1202 g
Dimensions: 254 x 203 x 37 mm
You may also be interested in...
£52.99Mixed media product
Please sign in to write a review
Simply reserve online and pay at the counter when you collect. Available in shop from just two hours, subject to availability.
Thank you for your reservation
Your order is now being processed and we have sent a confirmation email to you at
When will my order be ready to collect?
Following the initial email, you will be contacted by the shop to confirm that your item is available for collection.
Call us on or send us an email at
Unfortunately there has been a problem with your order
Please try again or alternatively you can contact your chosen shop on or send us an email at