Data 102: Inference
Data102 Notes #
Here are my notes for the Fall 2022 offering of Data 102, Berkeley’s Inference for Data Science course.
Data 102 explores two major concepts: making decisions under uncertainty and modeling the real world. This is all about making assumptions– here are some definitions:
- Frequentist: (data) is random, (parameter) is fixed
- Bayesian: is random, is random
- Parametric: Make assumptions about the relationship between and , then use these assumptions to find the best value of given
- Nonparametric: Don’t make any assumptions, and find any good function such that
Table of Contents #
- Binary Decision Making
Binary Decision Making is the simplest kind of decision we can make: 1 or 0, yes or now, true or...
- Hypothesis Testing: Null/alternative hypotheses, multiple hypothesis testing, controlling FWER/FDR, online decision making, likelihood ratios
- Decision Theory: Loss functions, risk, bias-variance tradeoff
- Parameter Estimation
Suppose we observe data points ( to ). Let be some unknown parameter that describes the distribution the data...
- Sampling
Intro In practice, getting the exact probability of an inference is not required as long as we get a rough estimate...
- Regression and GLMs
Regression and GLMs
Posterior Predictive Distribution Posterior Predictive Distribution: "if we saw some data, what future data might we expect?" = ...
- Nonparametric Methods
What does nonparametric mean? Nonparametric methods make no assumptions about the distribution of the data or parameters; the null hypothesis is...
- Interpretability
What do we look for in predictions? Accuracy: We want predictions to be close to the true values. Simplicity: We want the...
- Causality
Prediction vs Causality Prediction: using data, can we guess what will be? Causation: does X cause y to...
- Concentration Inequalities
The goal of concentration inequalities is to provide bounds on the probability of a random variable taking values in its...
- Bandits
Main idea: making repeated decisions based on feedback, factoring in the tradeoff between exploring new decisions or keeping existing good...
- Markov Decision Processes
What is a Markov Decision Process? A Markov Decision Process is a Markov model that solves nondeterministic search problems (where an...
- Reinforcement Learning
Introduction Reinforcement Learning (RL) is an example of online planning, where agents have no prior knowledge of rewards or transitions and...
How to contribute #
See the
contributing guide
Thanks for your interest in contributing to my notes! There's a lot of room for improvement, and I don't have...Contributing
For the most part, these notes should be pretty complete in terms of content, but could use some cleaning up (as well as more examples).