How to Read an Empirical Paper
A systematic framework for reading causal inference papers critically. Learn to identify the estimand, evaluate identification strategies, and judge credibility.
Research practices, workflows, and practical skills for credible empirical work.
Statistical methods and procedures organized by where they fit in your research pipeline.
Commit to your analysis before seeing the results — the antidote to the garden of forking paths.
How large a sample do you need to detect your effect? Power calculations prevent underpowered studies.
When conventional asymptotics fail — few clusters, unusual randomization — Fisher's exact approach provides valid inference.
When to cluster standard errors, at what level, and what to do when you have few clusters.
Testing many hypotheses inflates false positives. Bonferroni, Holm, BH-FDR, and Romano-Wolf corrections.
How robust are your results to omitted variable bias? Oster (2019) and Cinelli & Hazlett (2020) provide formal answers.
How much do your results depend on the specific analytical choices you made? Explore the full space of defensible specifications.
When point identification fails — especially due to differential attrition — informative bounds can still be useful.
Hands-on walkthroughs for reading papers, writing results, replicating studies, choosing methods, and avoiding common pitfalls.
A systematic framework for reading causal inference papers critically. Learn to identify the estimand, evaluate identification strategies, and judge credibility.
When and how to generalize causal estimates beyond the study population. PATE vs SATE vs LATE, site-selection bias, reweighting for target populations.
How to present causal inference results in academic papers. Covers table formatting, effect size reporting, and common writing mistakes.
Decision tree for identifying and exploiting natural experiments. From finding variation to choosing estimators, running diagnostics, and reporting results.
Workflow for causal inference from observational data. Covers estimator selection, managing the assumption burden, and building credibility without experiments.
When to use heteroscedasticity-robust, clustered, two-way clustered, Conley spatial, or wild bootstrap standard errors. Decision tree with code for every SE type in R, Python, and Stata.
A practical comparison of Difference-in-Differences and Synthetic Control methods, covering assumptions, strengths, and when each is appropriate. Includes the Synthetic DiD hybrid.
A practical comparison of selection-on-observables estimators: matching, inverse probability weighting, and doubly robust/AIPW methods. Covers assumptions, tradeoffs, and when to use each.
A practical guide to choosing between OLS regression with controls and Double/Debiased Machine Learning. Covers high-dimensional confounders, flexible functional forms, and regularization bias.
Step-by-step guide for replicating published causal inference papers. Covers finding data, reproducing results, diagnosing discrepancies, and extending analyses.
Common mistakes and anti-patterns to avoid in research design, estimation, interpretation, and reporting, with explanations and fixes.