My theory, which is mine
The premise of this book, and the other books in the Think X series, is that if you know how to program, you can use that skill to learn other topics.
Most books on Bayesian statistics use mathematical notation and present ideas in terms of mathematical concepts like calculus. This book uses Python code instead of math, and discrete approximations instead of continuous mathematics. As a result, what would be an integral in a math book becomes a summation, and most operations on probability distributions are simple loops.
I think this presentation is easier to understand, at least for people with programming skills. It is also more general, because when we make modeling decisions, we can choose the most appropriate model without worrying too much about whether the model lends itself to conventional analysis.
Also, it provides a smooth development path from simple examples to real-world problems. ChapterÂ 3 is a good example. It starts with a simple example involving dice, one of the staples of basic probability. From there it proceeds in small steps to the locomotive problem, which I borrowed from Mostellerâs Fifty Challenging Problems in Probability with Solutions, and from there to the German tank problem, a famously successful application of Bayesian methods during World War II.
Modeling and approximation
Most chapters in this book are motivated by a real-world problem, so they involve some degree of modeling. Before we can apply Bayesian methods (or any other analysis), we have to make decisions about which parts of the real-world system to include in the model and which details we can abstract away.
For example, in ChapterÂ 7, the motivating problem is to predict the winner of a hockey game. I model goal-scoring as a Poisson process, which implies that a goal is equally likely at any point in the game. That is not exactly true, but it is probably a good enough model for most purposes.
In ChapterÂ 12 the motivating problem is interpreting SAT scores (the SAT is a standardized test used for college admissions in the United States). I start with a simple model that assumes that all SAT questions are equally difficult, but in fact the designers of the SAT deliberately include some questions that are relatively easy and some that are relatively hard. I present a second model that accounts for this aspect of the design, and show that it doesnât have a big effect on the results after all.
I think it is important to include modeling as an explicit part of problem solving because it reminds us to think about modeling errors (that is, errors due to simplifications and assumptions of the model).
Many of the methods in this book are based on discrete distributions, which makes some people worry about numerical errors. But for real-world problems, numerical errors are almost always smaller than modeling errors.
Furthermore, the discrete approach often allows better modeling decisions, and I would rather have an approximate solution to a good model than an exact solution to a bad model.
On the other hand, continuous methods sometimes yield performance advantagesâfor example by replacing a linear- or quadratic-time computation with a constant-time solution.
So I recommend a general process with these steps:
While you are exploring a problem, start with simple models and implement them in code that is clear, readable, and demonstrably correct. Focus your attention on good modeling decisions, not optimization.
Once you have a simple model working, identify the biggest sources of error. You might need to increase the number of values in a discrete approximation, or increase the number of iterations in a Monte Carlo simulation, or add details to the model.
If the performance of your solution is good enough for your application, you might not have to do any optimization. But if you do, there are two approaches to consider. You can review your code and look for optimizations; for example, if you cache previously computed results you might be able to avoid redundant computation. Or you can look for analytic methods that yield computational shortcuts.
One benefit of this process is that Steps 1 and 2 tend to be fast, so you can explore several alternative models before investing heavily in any of them.
Another benefit is that if you get to Step 3, you will be starting with a reference implementation that is likely to be correct, which you can use for regression testing (that is, checking that the optimized code yields the same results, at least approximately).
Working with the code
The code and sound samples used in this book are available from https://github.com/AllenDowney/ThinkBayes. Git is a version control system that allows you to keep track of the files that make up a project. A collection of files under Gitâs control is called a ârepositoryâ. GitHub is a hosting service that provides storage for Git repositories and a convenient web interface.
The GitHub homepage for my repository provides several ways to work with the code:
You can create a copy of my repository on GitHub by pressing the Fork button. If you donât already have a GitHub account, youâll need to create one. After forking, youâll have your own repository on GitHub that you can use to keep track of code you write while working on this book. Then you can clone the repo, which means that you copy the files to your computer.
If you donât want to use Git at all, you can download the files in a Zip file using the button in the lower-right corner of the GitHub page.
The code for the first edition of the book works with Python 2. If you are using Python 3, you might want to use the updated code in https://github.com/AllenDowney/ThinkBayes2 instead.
I developed this book using Anaconda from Continuum Analytics, which is a free Python distribution that includes all the packages youâll need to run the code (and lots more). I found Anaconda easy to install. By default it does a user-level installation, not system-level, so you donât need administrative privileges. You can download Anaconda from http://continuum.io/downloads.
If you donât want to use Anaconda, you will need the following packages:
NumPy for basic numerical computation, http://www.numpy.org/;
SciPy for scientific computation, http://www.scipy.org/;
matplotlib for visualization, http://matplotlib.org/.
Although these are commonly used packages, they are not included with all Python installations, and they can be hard to install in some environments. If you have trouble installing them, I recommend using Anaconda or one of the other Python distributions that include these packages.
Many of the examples in this book use classes and functions defined in
thinkbayes.py. Some of them also use
thinkplot.py, which provides wrappers for some of the functions in
pyplot, which is part of
Experienced Python programmers will notice that the code in this book does not comply with PEP 8, which is the most common style guide for Python (http://www.python.org/dev/peps/pep-0008/).
Specifically, PEP 8 calls for lowercase function names with
underscores between words,
like_this. In this book and the accompanying code,
function and method names begin with a capital letter and use camel case,
I broke this rule because I developed some of the code while I was a Visiting Scientist at Google, so I followed the Google style guide, which deviates from PEP 8 in a few places. Once I got used to Google style, I found that I liked it. And at this point, it would be too much trouble to change.
Also on the topic of style, I write âBayesâs theoremâ with an s after the apostrophe, which is preferred in some style guides and deprecated in others. I donât have a strong preference. I had to choose one, and this is the one I chose.
And finally one typographical note: throughout the book, I use PMF and CDF for the mathematical concept of a probability mass function or cumulative distribution function, and Pmf and Cdf to refer to the Python objects I use to represent them.
There are several excellent modules for doing Bayesian statistics in
pymc and OpenBUGS. I
chose not to use them for this book because you need a fair amount of
background knowledge to get started with these modules, and I want to keep
the prerequisites minimal. If you know Python and a little bit about
probability, you are ready to start this book.
ChapterÂ 1 is about probability and Bayesâs theorem;
it has no code. ChapterÂ 2 introduces
Pmf, a thinly disguised Python dictionary I use
to represent a probability mass function (PMF). Then ChapterÂ 3 introduces
Suite, a kind of Pmf that provides a framework
for doing Bayesian updates. And thatâs just about all there is to
Well, almost. In some of the later chapters, I use analytic distributions including the Gaussian (normal) distribution, the exponential and Poisson distributions, and the beta distribution. In ChapterÂ 15 I break out the less-common Dirichlet distribution, but I explain it as I go along. If you are not familiar with these distributions, you can read about them on Wikipedia. You could also read the companion to this book, Think Stats, or an introductory statistics book (although Iâm afraid most of them take a mathematical approach that is not particularly helpful for practical purposes).
Conventions Used in This Book
The following typographical conventions are used in this book:
Indicates new terms, URLs, email addresses, filenames, and file extensions.
Used for program listings, as well as within paragraphs to refer to program elements such as variable or function names, databases, data types, environment variables, statements, and keywords.
Constant width bold
Shows commands or other text that should be typed literally by the user.
Constant width italic
Shows text that should be replaced with user-supplied values or by values determined by context.
This icon signifies a tip, suggestion, or general note.
This icon indicates a warning or caution.
SafariÂ® Books Online
Technology professionals, software developers, web designers, and business and creative professionals use Safari Books Online as their primary resource for research, problem solving, learning, and certification training.
Safari Books Online offers a range of plans and pricing for enterprise, government, education, and individuals. Members have access to thousands of books, training videos, and prepublication manuscripts in one fully searchable database from publishers like OâReilly Media, Prentice Hall Professional, Addison-Wesley Professional, Microsoft Press, Sams, Que, Peachpit Press, Focal Press, Cisco Press, John Wiley & Sons, Syngress, Morgan Kaufmann, IBM Redbooks, Packt, Adobe Press, FT Press, Apress, Manning, New Riders, McGraw-Hill, Jones & Bartlett, Course Technology, and hundreds more. For more information about Safari Books Online, please visit us online.
How to Contact Us
Please address comments and questions concerning this book to the publisher:
- OâReilly Media, Inc.
- 1005 Gravenstein Highway North
- Sebastopol, CA 95472
- 800-998-9938 (in the United States or Canada)
- 707-829-0515 (international or local)
- 707-829-0104 (fax)
We have a web page for this book, where we list errata, examples, and any additional information. You can access this page at http://oreil.ly/think-bayes.
To comment or ask technical questions about this book, send email to firstname.lastname@example.org.
For more information about our books, courses, conferences, and news, see our website at http://www.oreilly.com.
Find us on Facebook: http://facebook.com/oreilly
Follow us on Twitter: http://twitter.com/oreillymedia
Watch us on YouTube: http://www.youtube.com/oreillymedia
If you include at least part of the sentence the error appears in, that makes it easy for me to search. Page and section numbers are fine, too, but not as easy to work with. Thanks!
First, I have to acknowledge David MacKayâs excellent book, Information Theory, Inference, and Learning Algorithms, which is where I first came to understand Bayesian methods. With his permission, I use several problems from his book as examples.
This book also benefited from my interactions with Sanjoy Mahajan, especially in fall 2012, when I audited his class on Bayesian Inference at Olin College.
I wrote parts of this book during project nights with the Boston Python User Group, so I would like to thank them for their company and pizza.
Jonathan Edwards sent in the first typo.
George Purkins found a markup error.
Olivier Yiptong sent several helpful suggestions.
Yuriy Pasichnyk found several errors.
Kristopher Overholt sent a long list of corrections and suggestions.
Robert Marcus found a misplaced i.
Max Hailperin suggested a clarification in ChapterÂ 1.
Markus Dobler pointed out that drawing cookies from a bowl with replacement is an unrealistic scenario.
Tom Pollard and Paul A. Giannaros spotted a version problem with some of the numbers in the train example.
Ram Limbu found a typo and suggested a clarification.
In spring 2013, students in my class, Computational Bayesian Statistics, made many helpful corrections and suggestions: Kai Austin, Claire Barnes, Kari Bender, Rachel Boy, Kat Mendoza, Arjun Iyer, Ben Kroop, Nathan Lintz, Kyle McConnaughay, Alec Radford, Brendan Ritter, and Evan Simpson.
Greg Marra and Matt Aasted helped me clarify the discussion of The Price is Right problem.
Marcus Ogren pointed out that the original statement of the locomotive problem was ambiguous.
Jasmine Kwityn and Dan Fauxsmith at OâReilly Media proofread the book and found many opportunities for improvement.