Data 102: Inference
Data102 Notes #
Here are my notes for the Fall 2022 offering of Data 102, Berkeley’s Inference for Data Science course.
Data 102 explores two major concepts: making decisions under uncertainty and modeling the real world. This is all about making assumptions– here are some definitions:
- Frequentist: (data) is random, (parameter) is fixed
- Bayesian: is random, is random
- Parametric: Make assumptions about the relationship between and , then use these assumptions to find the best value of given
- Nonparametric: Don’t make any assumptions, and find any good function such that
Table of Contents #
- Binary Decision Making: Confusion matrix, sensitivity, specificity, TPR/FPR/FNR/FDP etc.
Binary Decision Making is the simplest kind of decision we can make: 1 or 0, yes or now, true or...
- Hypothesis Testing: Null/alternative hypotheses, multiple hypothesis testing, controlling FWER/FDR, online decision making, likelihood ratios
- Decision Theory: Loss functions, risk, bias-variance tradeoff
- Parameter Estimation: Likelihood, MLE, Bayesian parameter estimation, Bayesian hierarchical models
Suppose we observe data points ( to ). Let be some unknown parameter that describes the distribution the data...
- Sampling: Markov chains, MCMC, Metropolis-Hastings, Rejection sampling, Gibbs sampling
Intro In practice, getting the exact probability of an inference is not required as long as we get a rough estimate...
- Regression and GLMs: Generalized linear models, posterior predictive check
Regression and GLMs
Posterior Predictive Distribution Posterior Predictive Distribution: "if we saw some data, what future data might we expect?" = ...
- Nonparametric Methods: K-Nearest Neighbors, decision trees, random forests, neural networks, gradient ascent/descent
What does nonparametric mean? Nonparametric methods make no assumptions about the distribution of the data or parameters; the null hypothesis is...
- Interpretability: Interpretability, explainability
What do we look for in predictions? Accuracy: We want predictions to be close to the true values. Simplicity: We want the...
- Causality: Colliders, confounders, structural causal models, risk ratios, potential outcomes framework
Prediction vs Causality Prediction: using data, can we guess what will be? Causation: does X cause y to...
- Concentration Inequalities: Markov, Chebychev, Chernoff, Hoeffding
The goal of concentration inequalities is to provide bounds on the probability of a random variable taking values in its...
- Bandits: Multi-Armed Bandit Framework, UCB/ETC, Thomson Sampling, Regret
Main idea: making repeated decisions based on feedback, factoring in the tradeoff between exploring new decisions or keeping existing good...
- Markov Decision Processes: Value iteration, Q-value iteration, Policy iteration
What is a Markov Decision Process? A Markov Decision Process is a Markov model that solves nondeterministic search problems (where an...
- Reinforcement Learning: Q-Learning
Introduction Reinforcement Learning (RL) is an example of online planning, where agents have no prior knowledge of rewards or transitions and...
How to contribute #
See the
contributing guide
Thanks for your interest in contributing to my notes! There's a lot of room for improvement, and I don't have...Contributing
For the most part, these notes should be pretty complete in terms of content, but could use some cleaning up (as well as more examples).