Book description
Discover valuable machine learning techniques you can understand and apply using just high-school math.In Grokking Machine Learning you will learn:
- Supervised algorithms for classifying and splitting data
- Methods for cleaning and simplifying data
- Machine learning packages and tools
- Neural networks and ensemble methods for complex datasets
Grokking Machine Learning teaches you how to apply ML to your projects using only standard Python code and high school-level math. No specialist knowledge is required to tackle the hands-on exercises using Python and readily available machine learning tools. Packed with easy-to-follow Python-based exercises and mini-projects, this book sets you on the path to becoming a machine learning expert.
About the Technology
Discover powerful machine learning techniques you can understand and apply using only high school math! Put simply, machine learning is a set of techniques for data analysis based on algorithms that deliver better results as you give them more data. ML powers many cutting-edge technologies, such as recommendation systems, facial recognition software, smart speakers, and even self-driving cars. This unique book introduces the core concepts of machine learning, using relatable examples, engaging exercises, and crisp illustrations.
About the Book
Grokking Machine Learning presents machine learning algorithms and techniques in a way that anyone can understand. This book skips the confused academic jargon and offers clear explanations that require only basic algebra. As you go, you’ll build interesting projects with Python, including models for spam detection and image recognition. You’ll also pick up practical skills for cleaning and preparing data.
What's Inside
- Supervised algorithms for classifying and splitting data
- Methods for cleaning and simplifying data
- Machine learning packages and tools
- Neural networks and ensemble methods for complex datasets
About the Reader
For readers who know basic Python. No machine learning knowledge necessary.
About the Author
Luis G. Serrano is a research scientist in quantum artificial intelligence. Previously, he was a Machine Learning Engineer at Google and Lead Artificial Intelligence Educator at Apple.
Quotes
Did you think machine learning is complicated and hard to master? It’s not! Read this book! Serrano demystifies some of the best-held secrets of the machine learning society.
- Sebastian Thrun, Founder, Udacity
The first step to take on your machine learning journey.
- Millad Dagdoni, Norwegian Labour and Welfare Administration
A nicely written guided introduction, especially for those who want to code but feel shaky in their mathematics.
- Erik D. Sapper, California Polytechnic State University
The most approachable introduction to machine learning I’ve had the pleasure to read in recent years. Highly recommended.
- Kay Engelhardt, devstats
Publisher resources
Table of contents
- Grokking Machine Learning
- inside front cover
- Copyright
- contents
- front matter
- 1 What is machine learning? It is common sense, except done by a computer
- 2 Types of machine learning
-
3 Drawing a line close to our points: Linear regression
- The problem: We need to predict the price of a house
- The solution: Building a regression model for housing prices
- How to get the computer to draw this line: The linear regression algorithm
- How do we measure our results? The error function
- Real-life application: Using Turi Create to predict housing prices in India
- What if the data is not in a line? Polynomial regression
- Parameters and hyperparameters
- Applications of regression
- Summary
- Exercises
-
4 Optimizing the training process: Underfitting, overfitting, testing, and regularization
- An example of underfitting and overfitting using polynomial regression
- How do we get the computer to pick the right model? By testing
- Where did we break the golden rule, and how do we fix it? The validation set
- A numerical way to decide how complex our model should be: The model complexity graph
- Another alternative to avoiding overfitting: Regularization
- Polynomial regression, testing, and regularization with Turi Create
- Summary
- Exercises
- 5 Using lines to split our points: The perceptron algorithm
-
6 A continuous approach to splitting points: Logistic classifiers
- Logistic classifiers: A continuous version of perceptron classifiers
- How to find a good logistic classifier? The logistic regression algorithm
- Coding the logistic regression algorithm
- Real-life application: Classifying IMDB reviews with Turi Create
- Classifying into multiple classes: The softmax function
- Summary
- Exercises
- 7 How do you measure classification models? Accuracy and its friends
- 8 Using probability to its maximum: The naive Bayes model
-
9 Splitting data by asking questions: Decision trees
- The problem: We need to recommend apps to users according to what they are likely to download
- The solution: Building an app-recommendation system
- Beyond questions like yes/no
- The graphical boundary of decision trees
- Real-life application: Modeling student admissions with Scikit-Learn
- Decision trees for regression
- Applications
- Summary
- Exercises
- 10 Combining building blocks to gain more power: Neural networks
- 11 Finding boundaries with style: Support vector machines and the kernel method
-
12 Combining models to maximize results: Ensemble learning
- With a little help from our friends
- Bagging: Joining some weak learners randomly to build a strong learner
- AdaBoost: Joining weak learners in a clever way to build a strong learner
- Gradient boosting: Using decision trees to build strong learners
- XGBoost: An extreme way to do gradient boosting
- Applications of ensemble methods
- Summary
- Exercises
-
13 Putting it all in practice: A real-life example of data engineering and machine learning
- The Titanic dataset
- Cleaning up our dataset: Missing values and how to deal with them
- Feature engineering: Transforming the features in our dataset before training the models
- Training our models
- Tuning the hyperparameters to find the best model: Grid search
- Using K-fold cross-validation to reuse our data as training and validation
- Summary
- Exercises
-
Appendix A. Solutions to the exercises
- Chapter 2: Types of machine learning
- Chapter 3: Drawing a line close to our points: Linear regression
- Chapter 4: Optimizing the training process: Underfitting, overfitting, testing, and regularization
- Chapter 5: Using lines to split our points: The perceptron algorithm
- Chapter 6: A continuous approach to splitting points: Logistic classifiers
- Chapter 7: How do you measure classification models? Accuracy and its friends
- Chapter 8: Using probability to its maximum: The naive Bayes model
- Chapter 9: Splitting data by asking questions: Decision trees
- Chapter 10: Combining building blocks to gain more power: Neural networks
- Chapter 11: Finding boundaries with style: Support vector machines and the kernel method
- Chapter 12: Combining models to maximize results: Ensemble learning
- Chapter 13: Putting it all in practice: A real-life example of data engineering and machine learning
- Appendix B. The math behind gradient descent: Coming down a mountain using derivatives and slopes
- Appendix C. References
- index
Product information
- Title: Grokking Machine Learning
- Author(s):
- Release date: December 2021
- Publisher(s): Manning Publications
- ISBN: 9781617295911
You might also like
book
Grokking Deep Learning
Grokking Deep Learning teaches you to build deep learning neural networks from scratch! In his engaging …
book
Hands-On Machine Learning with Scikit-Learn, Keras, and TensorFlow, 2nd Edition
Through a series of recent breakthroughs, deep learning has boosted the entire field of machine learning. …
book
AI and Machine Learning for Coders
If you're looking to make a career move from programmer to AI specialist, this is the …
book
Designing Machine Learning Systems
Machine learning systems are both complex and unique. Complex because they consist of many different components …