82,99 €
inkl. MwSt.
Versandkostenfrei*
Versandfertig in 6-10 Tagen
  • Gebundenes Buch

Machine learning methods are now an important tool for scientists, researchers, engineers and students in a wide range of areas. This book is written for people who want to adopt and use the main tools of machine learning, but aren't necessarily going to want to be machine learning researchers. Intended for students in final year undergraduate or first year graduate computer science programs in machine learning, this textbook is a machine learning toolkit. Applied Machine Learning covers many topics for people who want to use machine learning processes to get things done, with a strong…mehr

Produktbeschreibung
Machine learning methods are now an important tool for scientists, researchers, engineers and students in a wide range of areas. This book is written for people who want to adopt and use the main tools of machine learning, but aren't necessarily going to want to be machine learning researchers. Intended for students in final year undergraduate or first year graduate computer science programs in machine learning, this textbook is a machine learning toolkit. Applied Machine Learning covers many topics for people who want to use machine learning processes to get things done, with a strong emphasis on using existing tools and packages, rather than writing one's own code.
A companion to the author's Probability and Statistics for Computer Science, this book picks up where the earlier book left off (but also supplies a summary of probability that the reader can use).

Emphasizing the usefulness ofstandard machinery from applied statistics, this textbook gives an overview of the major applied areas in learning, including coverage of:- classification using standard machinery (naive bayes; nearest neighbor; SVM)- clustering and vector quantization (largely as in PSCS)- PCA (largely as in PSCS)- variants of PCA (NIPALS; latent semantic analysis; canonical correlation analysis)- linear regression (largely as in PSCS)- generalized linear models including logistic regression- model selection with Lasso, elasticnet- robustness and m-estimators- Markov chains and HMM's (largely as in PSCS)- EM in fairly gory detail; long experience teaching this suggests one detailed example is required, which students hate; but once they've been through that, the next one is easy- simple graphical models (in the variational inference section)- classification with neural networks, with a particular emphasis onimage classification- autoencoding with neural networks- structure learning
Autorenporträt
David Forsyth grew up in Cape Town. He received a B.Sc. (Elec. Eng.) from the University of the Witwatersrand, Johannesburg in 1984, an M.Sc. (Elec. Eng.) from that university in 1986, and a D.Phil. from Balliol College, Oxford in 1989. He spent three years on the faculty at the University of Iowa, ten years on the faculty at the University of California at Berkeley, and then moved to the University of Illinois. He served as program co-chair for IEEE Computer Vision and Pattern Recognition in 2000, 2011, 2018 and 2021; general co-chair for CVPR 2006 and ICCV 2019, and program co-chair for the European Conference on Computer Vision 2008, and is a regular member of the program committee of all major international conferences on computer vision. He has served six terms on the SIGGRAPH program committee. In 2006, he received an IEEE technical achievement award, in 2009 he was named an IEEE Fellow, and in 2014 he was named an ACM Fellow. He served as Editor-in-Chief of IEEE TPAMI from 2014-2017. He is lead co-author of  Computer Vision: A Modern Approach, a textbook of computer vision that ran to two editions and four languages. He is sole author of Probability and Statistics for Computer Science, which provides the background for this book. Among a variety of odd hobbies, he is a compulsive diver, certi¿ed up to normoxic trimix level.