Event Studies (Dynamic Treatment Effects)
Visualize how treatment effects evolve over time — and test whether pre-trends support the parallel trends assumption.
Quick Reference
- When to Use
- When you want to visualize the time path of treatment effects and test for pre-existing differential trends. An event study is the dynamic generalization of DiD.
- Key Assumption
- Parallel trends in the pre-period (no differential trends before treatment) and no anticipation effects. Pre-treatment coefficients near zero support but do not prove this assumption.
- Common Mistake
- Interpreting non-significant pre-trend coefficients as proof that parallel trends holds (absence of evidence is not evidence of absence). Also, using standard TWFE event studies with staggered treatment timing without checking for heterogeneity bias.
- Estimated Time
- 2.5 hours
One-Line Implementation
reghdfe y ib(-1).event_time, absorb(unit_id year) vce(cluster state)feols(y ~ i(event_time, ref = -1) | unit_id + year, data = df, vcov = ~state)pf.feols('y ~ i(event_time, ref=-1) | unit + year', data=df, vcov={'CRV1': 'state'})Download Full Analysis Code
Complete scripts with diagnostics, robustness checks, and result export.
Motivating Example
A publicly traded firm announces that its longtime CEO is stepping down and a new CEO will take over. You want to know: what is the causal effect of CEO succession on stock prices?
You cannot just compare stock prices before and after the announcement — maybe the whole market was going up, or maybe the firm was already in decline. A simple before-after comparison confounds the succession effect with everything else happening at the same time.
What you need is a picture. You want to see what happened to the stock price in the weeks before the announcement (to check whether something was already going on) and in the weeks after (to see the actual effect and how it evolves). Did the stock jump immediately? Did it drift? Did the effect fade or grow?
That time-profile picture is an event study.
A. Overview: What Is an Event Study?
An event study is a regression framework that estimates separate treatment effects for each time period relative to the event. Instead of getting one number ("the treatment effect is 3.2 percentage points"), you get a time profile: the effect at the moment of treatment, one period after, two periods after, and so on — plus coefficients for the periods before treatment that serve as a diagnostic check.
The core idea is elegantly simple. You define event time as the number of periods relative to when the treatment happens. If a state adopts a policy in 2010, then for that state, event time -2 is 2008, event time 0 is 2010, and event time +3 is 2013. You then estimate a coefficient for each event-time indicator.
In finance, event studies have been used since at least Fama et al. (1969).
In the broader social sciences, event-study designs have become the standard way to visualize treatment dynamics in difference-in-differences settings. In many applied fields, an event-study plot is now a standard expectation in DiD papers.
The Static DiD Problem
Recall the canonical 2x2 DiD specification:
This specification gives you a single coefficient that averages the treatment effect across all post-treatment periods. But what if the effect takes three years to materialize? What if it fades after one year? What if it was already trending upward before treatment? The static model hides all of this dynamic information.
Common Confusions
"Which period should I omit as the reference?" The standard practice is to omit the period immediately before treatment (event time -1). This normalization ensures all coefficients are measured relative to the last pre-treatment period. Some researchers omit -2 instead, which is fine — just be consistent and transparent.
"Should I bin the endpoints?" If you have observations far from the event (e.g., event time -10 or +15), you typically bin them: group all observations at or below some threshold into a single indicator (e.g., event time ) and similarly for the right tail. This binning avoids estimating noisy coefficients from small samples at the extremes. But be careful — the binned coefficient averages effects across all those periods, which can be misleading if there are trends in the tails.
"Is an event study different from DiD?" An event study is a DiD. It is the dynamic generalization. The static DiD is a special case where you constrain all post-treatment effects to be the same and all pre-treatment effects to be zero.
B. Identification
The event-study specification is:
where and are unit and time fixed effects, is the event time for unit at calendar time , and the sum runs over all event-time indicators except the reference period .
Identifying assumptions:
-
Parallel trends: In the absence of treatment, treated and control units would have followed the same trajectory. Formally, the counterfactual trend in outcomes absent treatment is the same for treated and control groups: .
-
No anticipation: The treatment has no effect before it is actually implemented. This condition means for all .
The pre-treatment coefficients are your diagnostic: if they are jointly close to zero, this pattern is consistent with both parallel trends and no anticipation. If they show a clear trend, either parallel trends or no anticipation (or both) may be violated.
C. Visual Intuition
The classic event-study plot has event time on the x-axis and coefficient estimates (with 95% confidence intervals) on the y-axis. The reference period (typically ) is normalized to zero.
A good event study has the following pattern: pre-treatment coefficients hover around zero with confidence intervals that include zero, then there is a clear jump at that persists (or grows, or fades) in the post-treatment periods.
A problematic event study has the following pattern: pre-treatment coefficients trend upward (or downward), suggesting that treated units were already diverging from control units before the treatment. If you see such a pattern, your parallel trends assumption is in trouble.
Event Study Anatomy
Adjust the true treatment effect, the presence of pre-trends, and the noise level. Watch how the event-study plot changes. With pre-trends, can you still tell what the treatment effect is?
Why Event Study Design?
Panel DiD with 200 units (100 treated), 10 periods, treatment at t = 5. Anticipation = 0.0, dynamic growth = 0.0.
Estimation Results
| Estimator | β̂ | SE | 95% CI | Bias |
|---|---|---|---|---|
| Static DiDclosest | 2.073 | 0.197 | [1.69, 2.46] | +0.073 |
| Event Study (avg post) | 2.099 | 0.197 | [1.71, 2.48] | +0.099 |
| True β | 2.000 | — | — | — |
0 = no anticipation; 1 = strong anticipation (violates parallel trends)
0 = constant effect; positive = growing; negative = fading
Base treatment effect at time 0
Total units (half treated). Larger N → less sampling noise.
Why the difference?
Pre-trend coefficients are all statistically insignificant (close to zero), supporting the parallel trends assumption. This is the key diagnostic that event studies provide over static DiD. With constant treatment effects, the static DiD (β̂ = 2.073) closely matches the true ATT (2.000). Both approaches recover the treatment effect well; the event study additionally validates the parallel-trends assumption. Static DiD deviation from true ATT: 0.073. The event study reveals the full dynamic path of treatment effects.
D. Mathematical Derivation
Don't worry about the notation yet — here's what this means in words: Each event-study coefficient estimates the average difference between treated and control units at a specific time relative to treatment, after removing unit-specific and time-specific components.
Start with the potential outcomes framework. For each unit at time , the observed outcome is:
where indicates being periods after treatment. Under parallel trends:
Substituting and adding pre-treatment indicators as diagnostic terms:
Under correct specification, the OLS estimator for (for ) identifies the average treatment effect at event time :
For , the coefficients should be zero under the null of parallel trends and no anticipation. A joint F-test of all pre-treatment coefficients provides a formal test.
Important caveat ((Sun & Abraham, 2021)): When treatment timing is staggered, the OLS estimates of from the specification above can be contaminated. Already-treated units serve as controls for later-treated units, leading to negative weights on some group-time treatment effects. The interaction-weighted estimator of Sun and Abraham (2021) fixes this by estimating cohort-specific effects and then aggregating properly.
E. Implementation
library(fixest)
# Standard event study using fixest
# i(event_time, ref = -1) creates event-time dummies with k=-1 as reference
# Unit and year fixed effects are absorbed via | unit_id + year
# Standard errors are clustered at the state level
est <- feols(y ~ i(event_time, ref = -1) | unit_id + year,
data = df, vcov = ~state)
# Plot event-study coefficients with 95% confidence intervals
iplot(est, main = "Event Study: Treatment Effect Over Time")
# Sun & Abraham (2021) estimator for staggered treatment timing
# sunab(cohort_var, time_var) avoids contamination from already-treated units
# treatment_year = year each unit was first treated (Inf if never treated)
est_sa <- feols(y ~ sunab(treatment_year, year) | unit_id + year,
data = df, vcov = ~state)
iplot(est_sa)F. Diagnostics
-
Joint F-test of pre-treatment coefficients. Test whether are jointly zero. This joint test is more informative than eyeballing individual coefficients. In Stata:
testparmon the pre-treatment indicators. In R with fixest:wald(est, "event_time::-"). -
Visual inspection of pre-trends. Even if the F-test does not reject, look at the pattern. A clear monotonic trend that approaches zero at is concerning even if individual coefficients are not significant.
-
Confidence interval widths. If your pre-treatment confidence intervals are wide enough to include economically meaningful effects, your pre-trend test lacks power. You cannot distinguish "no pre-trend" from "pre-trend we cannot detect."
-
Placebo treatment dates. Artificially reassign treatment to earlier periods and re-estimate. If you still see "effects," something is wrong.
-
Sensitivity to endpoint binning. Re-estimate with different bin cutoffs. If your results change substantially, the specification is fragile.
Interpreting Your Results
Pre-treatment coefficients near zero: This finding supports (but does not prove) parallel trends. Report the joint F-test p-value alongside the visual.
Pre-treatment coefficients trending: This pattern is a red flag. Consider whether the trend is economically meaningful. If it is large relative to your post-treatment effects, your results are not credible without further adjustment (e.g., including unit-specific linear trends, though this introduces its own problems).
Post-treatment effects that grow over time: This dynamic is common and often substantively meaningful. A policy may take time to have its full effect. Report the dynamics, not just the average.
Post-treatment effects that shrink: This pattern could mean the effect is transitory, or that control units are "catching up." Think about the mechanism.
G. What Can Go Wrong
Staggered Timing Contaminates Event-Study Coefficients
Researcher uses the Sun and Abraham (2021) interaction-weighted estimator for an event study of state health-insurance expansions adopted between 2010 and 2018, with a never-treated control group.
Post-treatment coefficients show a gradual increase in insurance coverage: +1.2 pp at k=0, +2.8 pp at k=1, +4.1 pp at k=2, consistent with phased enrollment dynamics.
Underpowered Pre-Trend Test Creates False Confidence
Researcher estimates an event study of CEO turnovers on firm investment rates with 200 firms and 10 pre-treatment quarters. They report both the event-study plot and the power of the pre-trend test, noting that the 95% CI at each pre-period ranges from -1.5 to +1.5 percentage points.
The researcher acknowledges: 'Our pre-trend test cannot rule out differential pre-trends as large as 1.5 pp per quarter, which over 4 quarters could generate a cumulative bias of 6 pp — comparable to our estimated post-treatment effect of 4.2 pp. We therefore view the parallel trends assumption as plausible but not definitive.'
Endpoint Binning Masks Problematic Dynamics
Researcher studies the effect of plant closings on local unemployment using event time from -8 to +8 years. They bin endpoints at k <= -6 and k >= 6, and separately report the un-binned specification as a robustness check.
The binned specification shows a clean pattern. The un-binned specification reveals that coefficients at k = -7 and k = -8 are trending upward, suggesting potential anticipation or pre-existing decline. The researcher investigates and finds that some closings were preceded by gradual layoffs.
H. Practice
In an event-study regression, the coefficient at event time k = -3 is 0.02 with a standard error of 0.15. The coefficient at event time k = +2 is 0.08 with a standard error of 0.03. What should you conclude?
Reading an Event Study: Minimum Wage Increases and Teen Employment
A researcher estimates an event study of state-level minimum wage increases on teen employment rates. She plots coefficients for 4 periods before and 4 periods after the minimum wage change, normalizing the period just before the change (k = -1) to zero. The pre-period coefficients (k = -4 through k = -2) are all close to zero and not statistically significant. Post-period coefficients decline steadily from -0.01 at k = +1 to -0.04 at k = +4.
Read the analysis below carefully and identify the errors.
Select all errors you can find:
Read the paper summary below and write a brief referee critique (2-3 sentences) of the identification strategy.
Paper Summary
The authors study the effect of CEO succession on cumulative abnormal stock returns using an event-study design around 450 CEO transitions at S&P 500 firms from 2000-2020. They estimate abnormal returns using a market model with a 250-day estimation window and a [-10, +10] day event window around the announcement.
Key Table
| Event Window | CAR | t-stat |
|---|---|---|
| [-10, -2] | 0.3% | 0.82 |
| [-1, 0] | 1.8% | 3.41 |
| [+1, +5] | 0.6% | 1.12 |
| [+6, +10] | -0.4% | 0.73 |
Authors' Identification Claim
The near-zero pre-event CAR confirms that markets did not anticipate the succession, and the significant CAR at the announcement date identifies the causal effect of CEO change on firm value.
I. Swap-In: When to Use Something Else
- Canonical DiD: When a single treatment-effect estimate is sufficient and the time profile of effects is not of primary interest.
- Staggered DiD estimators: When the goal is a single aggregate ATT across cohorts — modern estimators (Callaway & Sant'Anna, Sun & Abraham) aggregate event-study coefficients into a robust summary.
- Synthetic control: When there is a single treated unit and constructing a counterfactual from donor units is more credible than assuming parallel trends.
- Interrupted time series (ITS): When there is no control group and the design relies entirely on modeling the pre-treatment trend to forecast the counterfactual.
J. Reviewer Checklist
Critical Reading Checklist
Paper Library
Foundational (9)
Fama, E. F., Fisher, L., Jensen, M. C., & Roll, R. (1969). The Adjustment of Stock Prices to New Information.
This paper is the origin of the modern event study methodology in finance. Fama, Fisher, Jensen, and Roll studied how stock prices adjust to stock splits and established the basic framework of measuring abnormal returns around corporate events that has been used in thousands of subsequent studies.
MacKinlay, A. C. (1997). Event Studies in Economics and Finance.
MacKinlay provided a comprehensive methodological survey of event studies, covering the statistical framework, estimation windows, abnormal return calculations, and testing procedures. This paper remains the standard reference for researchers designing and implementing event studies.
Brown, S. J., & Warner, J. B. (1985). Using Daily Stock Returns: The Case of Event Studies.
Brown and Warner extended the event study framework from monthly to daily stock returns and examined the statistical properties of various test statistics. Their simulations showed that simple methods perform well in most settings, providing practical reassurance for applied researchers.
Kothari, S. P., & Warner, J. B. (2007). Econometrics of Event Studies.
Kothari and Warner updated the survey of event study methods, covering long-horizon event studies, cross-sectional regression approaches, and the econometric challenges that arise with overlapping events and event-induced variance changes.
Borusyak, K., Jaravel, X., & Spiess, J. (2024). Revisiting Event-Study Designs: Robust and Efficient Estimation.
Borusyak, Jaravel, and Spiess proposed an imputation estimator for event studies that first estimates unit and time fixed effects from untreated observations, then imputes counterfactual outcomes for treated units. This approach is efficient, avoids negative weighting problems of two-way fixed effects, and is now a leading method for robust event-study estimation.
Rambachan, A., & Roth, J. (2023). A More Credible Approach to Parallel Trends.
Rambachan and Roth developed a sensitivity analysis framework for assessing the robustness of event-study and difference-in-differences estimates to violations of the parallel trends assumption. Their approach constructs honest confidence intervals under restrictions on how pre-trends can extrapolate into the post-treatment period, providing a disciplined alternative to informal pre-trend tests.
Freyaldenhoven, S., Hansen, C., & Shapiro, J. M. (2019). Pre-Event Trends in the Panel Event-Study Design.
Freyaldenhoven, Hansen, and Shapiro developed diagnostic tools and an instrumental-variables-based estimator for panel event-study designs when pre-event trends may be present. Their framework helps researchers distinguish true anticipation effects from confounding trends, addressing a central challenge in event-study credibility.
Sun, L., & Abraham, S. (2021). Estimating Dynamic Treatment Effects in Event Studies with Heterogeneous Treatment Effects.
The foundational paper for understanding why standard event-study regressions can break down under staggered treatment timing with heterogeneous effects. When already-treated units serve as controls for later-treated units, the TWFE event-study coefficients can be contaminated. Sun and Abraham propose the interaction-weighted estimator as a solution that properly aggregates cohort-specific effects.
Roth, J. (2022). Pretest with Caution: Event-Study Estimates after Testing for Parallel Trends.
Demonstrates that conditioning on passing a pre-trend test can distort post-treatment estimates. When researchers run a pre-trend test and proceed only if it passes, the resulting estimates are biased because the conditioning event is informative about the magnitude of any underlying trend. Proposes honest confidence intervals that account for pre-testing.
Application (4)
McWilliams, A., & Siegel, D. (1997). Event Studies in Management Research: Theoretical and Empirical Issues.
McWilliams and Siegel introduced event study methods to the management research community, explaining the assumptions, methodology, and common pitfalls. This tutorial article led to widespread adoption of event studies in strategic management research.
Koh, P.-S., & Reeb, D. M. (2015). Missing R&D.
Koh and Reeb used event study methodology to examine stock market reactions to firms that report versus do not report R&D expenditures, showing that missing R&D data is not random and has implications for how investors value innovation. This paper illustrates event study methods applied to accounting and disclosure questions.
Dranove, D., & Olsen, C. (1994). The Economic Side Effects of Dangerous Drug Announcements.
Dranove and Olsen used event studies to measure the stock market impact of FDA drug safety announcements on pharmaceutical firms. This application demonstrated how event studies can quantify the financial consequences of regulatory actions in health care and management contexts.
Capron, L., & Pistre, N. (2002). When Do Acquirers Earn Abnormal Returns?.
Capron and Pistre used event study methodology to examine when acquiring firms earn positive abnormal returns from mergers and acquisitions. They found that acquirers earn positive returns only when they are the primary source of value creation, contributing to the M&A strategy literature.
Survey (1)
de Chaisemartin, C., & D'Haultfoeuille, X. (2023). Two-Way Fixed Effects and Differences-in-Differences with Heterogeneous Treatment Effects: A Survey.
De Chaisemartin and D'Haultfoeuille provided a comprehensive survey of the recent literature on problems with two-way fixed effects estimators under heterogeneous treatment effects, covering the key diagnostic tests, alternative estimators, and practical guidance for applied researchers working with event-study and difference-in-differences designs.