Video description
9 Hours of Video InstructionHandson approach to learning the probability and statistics underlying machine learning
Overview
Probability and Statistics for Machine Learning (Machine Learning Foundations) LiveLessons provides you with a functional, handson understanding of probability theory and statistical modeling, with a focus on machine learning applications.
About the Instructor
Jon Krohn is Chief Data Scientist at the machine learning company untapt. He authored the book Deep Learning Illustrated, an instant #1 bestseller that has been translated into six languages. Jon is renowned for his compelling lectures, which he offers in person at Columbia University and New York University, as well as online via O'Reilly, YouTube, and the SuperDataScience podcast. Jon holds a PhD from Oxford and has been publishing on machine learning in leading academic journals since 2010; his papers have been cited over a thousand times.
Skill Level
 Intermediate
Learn How To
 Understand the appropriate variable type and probability distribution for representing a given class of data
 Calculate all of the standard summary metrics for describing probability distributions, as well as the standard techniques for assessing the relationships between distributions
 Apply information theory to quantify the proportion of valuable signal that's present among the noise of a given probability distribution
 Hypothesize about and critically evaluate the inputs and outputs of machine learning algorithms using essential statistical tools such as the ttest, ANOVA, and Rsquared
 Understand the fundamentals of both frequentist and Bayesian statistics, as well as appreciate when one of these approaches is appropriate for the problem you're solving
 Use historical data to predict the future using regression models that take advantage of frequentist statistical theory (for smaller data sets) and modern machine learning theory (for larger data sets), including why we may want to consider applying deep learning to a given problem
 Develop a deep understanding of what's going on beneath the hood of predictive statistical models and machine learning algorithms
Who Should Take This Course
 You use highlevel software libraries (e.g., scikitlearn, Keras, TensorFlow) to train or deploy machine learning algorithms and would now like to understand the fundamentals underlying the abstractions, enabling you to expand your capabilities
 You're a software developer who would like to develop a firm foundation for the deployment of machine learning algorithms into production systems
 You're a data scientist who would like to reinforce your understanding of the subjects at the core of your professional discipline
 You're a data analyst or AI enthusiast who would like to become a data scientist or data/ML engineer, and so you're keen to deeply understand the field you're entering from the ground up (very wise of you!)
Course Requirements
 Mathematics: Familiarity with secondary schoollevel mathematics will make it easier for you to follow along with the class. If you are comfortable dealing with quantitative informationsuch as understanding charts and rearranging simple equationsthen you should be wellprepared to follow along with all of the mathematics.
 Programming: All code demos will be in Python so experience with it or another objectoriented programming language would be helpful for following along with the handson examples.
Lesson Descriptions
Lesson 1: Introduction to Probability
In Lesson 1, Jon starts by orienting you to the machine learning foundations series and covering what probability theory is. He then begins coverage of the most essential probability concepts, which is reinforced by comprehension exercises. The lesson ends with a comparison of Bayesian and frequentist statistics, as well as a discussion of applications of probability to machine learning.
Lesson 2: Random Variables
Lesson 2 focuses on random variables, a fundamental probability concept that is a prerequisite for understanding the later lessons. Jon starts off with an exploration of discrete and continuous variables as well as the probability distributions to which they correspond. The lesson wraps up with calculation of the expected value of random variables.
Lesson 3: Describing Distributions
Lesson 3 is all about metrics for describing probability distributions. Jon covers measures of central tendency, quantiles, boxandwhisker plots, measures of dispersion, and measures of relatedness.
Lesson 4: Relationships Between Probabilities
In Lesson 4, Jon explores the core relationships between probabilities, including joint distributions, marginal and conditional probabilities, the chain rule, and independence.
Lesson 5: Distributions in Machine Learning
Having now led you through mastering probability theory in general, in Lesson 5 Jon details the most important probability distributions in machine learning, including the uniform and normal distributions, as well as the critical concept of the central limit theorem. He also covers the lognormal, exponential, discrete, and Poisson distributions, as well as mixtures of distributions and how to prepare distributions for input into a machine learning model.
Lesson 6: Information Theory
In Lesson 6, Jon provides you with an introduction to information theory, a field of study related to probability theory that includes some key concepts that are ubiquitous in machine learning. Specifically, he defines selfinformation, Shannon entropy, KL divergence, and crossentropy.
Lesson 7: Introduction to Statistics
From Lesson 7 onward, Jon shifts gears from general probability theory to the statistical models that probability theory facilitates. He starts by explaining how statistics are applied to machine learning and reviewing the most essential probability theory you absolutely must know to move forward. He then introduces new statistics concepts, specifically zscores and pvalues.
Lesson 8: Comparing Means
In Lesson 8, Jon teaches you to use probability and statistics to compare distributions with ttests. He covers all the critical types, including the singlesample, independent, and paired varieties. Jon provides specific applications of ttests to machine learning, and then wraps the lesson up with a discussion of related concepts, namely, confidence intervals and analysis of variance.
Lesson 9: Correlation
Lesson 9 builds on the introduction to correlation in Lesson 3. You are now armed with enough statistical knowledge to calculate pvalues for correlations and calculate the coefficient of determination. Jon finishes off the lesson with important discussions about inferring causation and correcting for multiple comparisons.
Lesson 10: Regression
You're in for a treat with Lesson 10, which brings together the preceding lessons with practical, realworld demonstrations of regressiona powerful, highly extensible approach to making predictions. Jon distinguishes independent from dependent variables and uses linear regression to predict continuous variablesfirst with a single model feature and then with many, including discrete features. The lesson concludes with logistic regression for predicting discrete outcomes.
Lesson 11: Bayesian Statistics
Lesson 11 is on Bayesian statistics. Jon provides a guide as to when frequentist statistics or Bayesian statistics might be the appropriate option for the problem you're solving. Jon then introduces the most essential Bayesian concepts. Finally, Jon leaves you with resources for studying probability and statistics beyond what we had time for in these LiveLessons themselves.
Notebooks are available at github.com/jonkrohn/MLfoundations
About Pearson Video Training
Pearson publishes expertled video tutorials covering a wide selection of technology topics designed to teach you the skills you need to succeed. These professional and personal technology videos feature worldleading author instructors published by your trusted technology brands: AddisonWesley, Cisco Press, Pearson IT Certification, Sams, and Que Topics include: IT Certification, Network Security, Cisco Technology, Programming, Web Development, Mobile Development, and more. Learn more about Pearson Video training at http://www.informit.com/video.
Table of contents
 Introduction

Lesson 1: Introduction to Probability
 Topics
 1.1 Orientation to the Machine Learning Foundations Series
 1.2 What Probability Theory Is
 1.3 Events and Sample Spaces
 1.4 Multiple Observations
 1.5 Factorials and Combinatorics
 1.6 Exercises
 1.7 The Law of Large Numbers and the Gambler's Fallacy
 1.8 Probability Distributions in Statistics
 1.9 Bayesian versus Frequentist Statistics
 1.10 Applications of Probability to Machine Learning
 Lesson 2: Random Variables
 Lesson 3: Describing Distributions
 Lesson 4: Relationships Between Probabilities
 Lesson 5: Distributions in Machine Learning
 Lesson 6: Information Theory
 Lesson 7: Introduction to Statistics
 Lesson 8: Comparing Means
 Lesson 9: Correlation

Lesson 10: Regression
 Topics
 10.1 Independent versus Dependent Variables
 10.2 Linear Regression to Predict Continuous Values
 10.3 Fitting a Line to Points on a Cartesian Plane
 10.4 Linear Least Squares Exercise
 10.5 Ordinary Least Squares
 10.6 Categorical "Dummy" Features
 10.7 Logistic Regression to Predict Categories
 10.8 OpenEnded Exercises
 Lesson 11: Bayesian Statistics
 Summary
Product information
 Title: Probability and Statistics for Machine Learning
 Author(s):
 Release date: May 2021
 Publisher(s): Pearson
 ISBN: 0137566271
You might also like
book
Statistics for Machine Learning
Build Machine Learning models with a sound statistical understanding. About This Book Learn about the statistics …
video
Probability / Statistics  The Foundations of Machine Learning
The objective of this course is to give you a solid foundation needed to excel in …
book
Machine Learning with PyTorch and ScikitLearn
This book of the bestselling and widely acclaimed Python Machine Learning series is a comprehensive guide …
book
Bayesian Statistics the Fun Way
Probability and statistics are increasingly important in a huge range of professions. But many people use …