MethodAtlas
Method·intermediate·12 min read
Design-BasedModern

Event Studies (Dynamic Treatment Effects)

Visualize how treatment effects evolve over time — and test whether pre-trends support the parallel trends assumption.

When to UseWhen you want to visualize the time path of treatment effects and test for pre-existing differential trends. An event study is the dynamic generalization of DiD.
AssumptionParallel trends in the pre-period (no differential trends before treatment) and no anticipation effects. Pre-treatment coefficients near zero support but do not prove this assumption.
MistakeInterpreting non-significant pre-trend coefficients as proof that parallel trends holds (absence of evidence is not evidence of absence). Also, using standard TWFE event studies with staggered treatment timing without checking for heterogeneity bias.
Reading Time~12 min read · 11 sections · 6 interactive exercises

One-Line Implementation

Rfeols(y ~ i(event_time, treated, ref = -1) | unit_id + year, data = df, vcov = ~state)
Statareghdfe y ib(-1).event_time#1.treated, absorb(unit_id year) vce(cluster state)
Pythonpf.feols('y ~ i(event_time, treated, ref=-1) | unit + year', data=df, vcov={'CRV1': 'state'})

Download Full Analysis Code

Complete scripts with diagnostics, robustness checks, and result export.

Motivating Example: CEO Succession and Stock Prices

A publicly traded firm announces that its longtime CEO is stepping down and a new CEO will take over. You want to know: what is the causal effect of CEO succession on stock prices?

You cannot just compare stock prices before and after the announcement — maybe the whole market was going up, or maybe the firm was already in decline. A simple before-after comparison confounds the succession effect with everything else happening at the same time.

What you need is a picture. You want to see what happened to the stock price in the weeks before the announcement (to check whether something was already going on) and in the weeks after (to see the actual effect and how it evolves). Did the stock jump immediately? Did it drift? Did the effect fade or grow?

That time-profile picture is an .


AOverview

An event study is a regression framework that estimates separate treatment effects for each time period relative to the event. Instead of getting one number ("the treatment effect is 3.2 percentage points"), you get a time profile: the effect at the moment of treatment, one period after, two periods after, and so on — plus coefficients for the periods before treatment that serve as a diagnostic check.

The core idea is elegantly simple. You define as the number of periods relative to when the treatment happens. If a state adopts a policy in 2010, then for that state, event time -2 is 2008, event time 0 is 2010, and event time +3 is 2013. You then estimate a coefficient for each event-time indicator.

In finance, event studies have been used since at least Fama et al. (1969).

In the broader social sciences, event-study designs have become the standard way to visualize treatment dynamics in difference-in-differences settings. In many applied fields, an event-study plot is now a standard expectation in Difference-in-Differences (DiD) papers.

The Static DiD Problem

Recall the canonical 2x2 DiD specification:

Yit=αi+λt+τDit+εitY_{it} = \alpha_i + \lambda_t + \tau \cdot D_{it} + \varepsilon_{it}

This specification gives you a single coefficient τ\tau that averages the treatment effect across all post-treatment periods. But what if the effect takes three years to materialize? What if it fades after one year? What if it was already trending upward before treatment? The static model hides all of this dynamic information.

Common Confusions

"Which period should I omit as the reference?" The standard practice is to omit the period immediately before treatment (event time -1). This normalization ensures all coefficients are measured relative to the last pre-treatment period. An alternative convention is to omit -2 instead, which is equally valid as long as the choice is consistent and transparent.

"Should I bin the endpoints?" If you have observations far from the event (e.g., event time -10 or +15), you typically bin them: group all observations at or below some threshold into a single indicator (e.g., event time 6\leq -6) and similarly for the right tail. This binning avoids estimating noisy coefficients from small samples at the extremes. But be careful — the binned coefficient averages effects across all those periods, which can be misleading if there are trends in the tails.

"Is an event study different from DiD?" In the panel/policy context, an event study is a dynamic generalization of DiD. The static DiD is a special case where you constrain all post-treatment effects to be the same and all pre-treatment effects to be zero.


BIdentification

The event-study specification is:

Yit=αi+λt+k1βk1[Kit=k]+εitY_{it} = \alpha_i + \lambda_t + \sum_{k \neq -1} \beta_k \cdot \mathbf{1}[K_{it} = k] + \varepsilon_{it}

where αi\alpha_i and λt\lambda_t are unit and time fixed effects, KitK_{it} is the event time for unit ii at calendar time tt, and the sum runs over all event-time indicators except the reference period k=1k = -1.

Identifying assumptions:

  1. Parallel trends: In the absence of treatment, treated and control units would have followed the same trajectory. Formally, the counterfactual trend in outcomes absent treatment is the same for treated and control groups: E[Yit(0)Yi,t1(0)Di=1]=E[Yit(0)Yi,t1(0)Di=0]E[Y_{it}(0) - Y_{i,t-1}(0) | D_i = 1] = E[Y_{it}(0) - Y_{i,t-1}(0) | D_i = 0].

  2. No anticipation: The treatment has no effect before it is actually implemented. Combined with parallel trends, this condition implies βk=0\beta_k = 0 for all k<0k < 0.

The pre-treatment coefficients {βk}k<1\{\beta_k\}_{k < -1} are your diagnostic: if they are jointly close to zero, this pattern is consistent with both parallel trends and no anticipation. If they show a clear trend, either parallel trends or no anticipation (or both) may be violated.


CVisual Intuition

The classic event-study plot has event time on the x-axis and coefficient estimates (with 95% confidence intervals) on the y-axis. The reference period (typically k=1k = -1) is normalized to zero.

A good event study has the following pattern: pre-treatment coefficients hover around zero with confidence intervals that include zero, then there is a clear jump at k=0k = 0 that persists (or grows, or fades) in the post-treatment periods.

A problematic event study has the following pattern: pre-treatment coefficients trend upward (or downward), suggesting that treated units were already diverging from control units before the treatment. If you see such a pattern, your parallel trends assumption is in trouble.


DMathematical Derivation

Don't worry about the notation yet — here's what this means in words: Each event-study coefficient estimates the average difference between treated and control units at a specific time relative to treatment, after removing unit-specific and time-specific components.

Start with the potential outcomes framework. For each unit ii at time tt, the observed outcome is:

Yit=Yit(0)+kβkDitkY_{it} = Y_{it}(0) + \sum_{k} \beta_k \cdot D_{it}^k

where Ditk=1[Kit=k,k0]D_{it}^k = \mathbf{1}[K_{it} = k, k \geq 0] indicates being kk periods after treatment. Under parallel trends:

E[Yit(0)i,t]=αi+λtE[Y_{it}(0) \mid i, t] = \alpha_i + \lambda_t

Substituting and adding pre-treatment indicators as diagnostic terms:

Yit=αi+λt+k1βk1[Kit=k]+εitY_{it} = \alpha_i + \lambda_t + \sum_{k \neq -1} \beta_k \cdot \mathbf{1}[K_{it} = k] + \varepsilon_{it}

Under correct specification, the OLS estimator for βk\beta_k (for k0k \geq 0) identifies the average treatment effect at event time kk:

β^kpE[Yit(1)Yit(0)Kit=k]\hat{\beta}_k \xrightarrow{p} E[Y_{it}(1) - Y_{it}(0) \mid K_{it} = k]

(under a single treatment date or homogeneous effects; with staggered adoption, OLS event-study coefficients can be contaminated by treatment effect heterogeneity — see Sun and Abraham (2021))

For k<1k < -1, the coefficients βk\beta_k should be zero under the null of parallel trends and no anticipation. A joint F-test of all pre-treatment coefficients provides a formal test.

Important caveat: When treatment timing is staggered, the OLS estimates of βk\beta_k from the specification above can be contaminated. Already-treated units serve as controls for later-treated units, leading to negative weights on some group-time treatment effects. The interaction-weighted estimator of Sun and Abraham (2021) fixes this by estimating cohort-specific effects and then aggregating properly.


EImplementation

# Requires: fixest
# fixest: fast fixed-effects estimation with multi-way clustering (Berge)
library(fixest)

# --- Step 1: Standard event study with TWFE ---
# i(event_time, ref = -1): creates event-time dummies, omitting k=-1 as reference
# | unit_id + year: absorbs unit and year fixed effects
# vcov = ~state: clusters SEs at state level (treatment assignment level)
est <- feols(y ~ i(event_time, ref = -1) | unit_id + year,
           data = df, vcov = ~state)

# iplot() displays coefficients with 95% confidence intervals
# Pre-treatment coefficients near zero support the parallel trends assumption
# Post-treatment coefficients show the dynamic treatment effect path
iplot(est, main = "Event Study: Treatment Effect Over Time")

# --- Step 2: Sun & Abraham (2021) estimator for staggered timing ---
# sunab() avoids contamination from using already-treated units as controls
# treatment_year = year each unit was first treated (Inf if never treated)
# Use this instead of standard TWFE when treatment timing varies across units
est_sa <- feols(y ~ sunab(treatment_year, year) | unit_id + year,
              data = df, vcov = ~state)
iplot(est_sa)
# Compare with Step 1: if estimates differ, treatment effect heterogeneity is present
Requiresfixest

FDiagnostics

  1. Joint F-test of pre-treatment coefficients. Test whether β2,β3,\beta_{-2}, \beta_{-3}, \ldots are jointly zero. This joint test is more informative than eyeballing individual coefficients. In Stata: testparm on the pre-treatment indicators. In R with fixest: wald(est, "event_time::-").

  2. Visual inspection of pre-trends. Even if the F-test does not reject, look at the pattern. A clear monotonic trend that approaches zero at k=1k = -1 is concerning even if individual coefficients are not significant.

  3. Confidence interval widths. If your pre-treatment confidence intervals are wide enough to include economically meaningful effects, your pre-trend test lacks power. You cannot distinguish "no pre-trend" from "pre-trend we cannot detect."

  4. Placebo treatment dates. Artificially reassign treatment to earlier periods and re-estimate. If you still see "effects," something is wrong.

  5. Sensitivity to endpoint binning. Re-estimate with different bin cutoffs. If your results change substantially, the specification is fragile.

Interpreting Your Results

Pre-treatment coefficients near zero: This finding supports (but does not prove) parallel trends. Report the joint F-test p-value alongside the visual.

Pre-treatment coefficients trending: This pattern is a red flag. Consider whether the trend is economically meaningful. If it is large relative to your post-treatment effects, your results are not credible without further adjustment (e.g., including unit-specific linear trends, though this introduces its own problems).

Post-treatment effects that grow over time: This dynamic is common and often substantively meaningful. A policy may take time to have its full effect. Report the dynamics, not just the average.

Post-treatment effects that shrink: This pattern could mean the effect is transitory, or that control units are "catching up." Think about the mechanism.


GWhat Can Go Wrong

What Can Go Wrong

Staggered Timing Contaminates Event-Study Coefficients

Researcher uses the Sun and Abraham (2021) interaction-weighted estimator for an event study of state health-insurance expansions adopted between 2010 and 2018, with a never-treated control group.

Post-treatment coefficients show a gradual increase in insurance coverage: +1.2 pp at k=0, +2.8 pp at k=1, +4.1 pp at k=2, consistent with phased enrollment dynamics.

What Can Go Wrong

Underpowered Pre-Trend Test Creates False Confidence

Researcher estimates an event study of CEO turnovers on firm investment rates with 200 firms and 10 pre-treatment quarters. They report both the event-study plot and the power of the pre-trend test, noting that the 95% CI at each pre-period ranges from -1.5 to +1.5 percentage points.

The researcher acknowledges: 'Our pre-trend test cannot rule out differential pre-trends as large as 1.5 pp per quarter, which over 4 quarters could generate a cumulative bias of 6 pp — comparable to our estimated post-treatment effect of 4.2 pp. We therefore view the parallel trends assumption as plausible but not definitive.'

What Can Go Wrong

Endpoint Binning Masks Problematic Dynamics

Researcher studies the effect of plant closings on local unemployment using event time from -8 to +8 years. They bin endpoints at k <= -6 and k >= 6, and separately report the un-binned specification as a robustness check.

The binned specification shows a clean pattern. The un-binned specification reveals that coefficients at k = -7 and k = -8 are trending upward, suggesting potential anticipation or pre-existing decline. The researcher investigates and finds that some closings were preceded by gradual layoffs.


HPractice

Concept Check

In an event-study regression, the coefficient at event time k = -3 is 0.02 with a standard error of 0.15. The coefficient at event time k = +2 is 0.08 with a standard error of 0.03. What should you conclude?

Concept Check

In an event study plot, what do the pre-treatment coefficient estimates test?

Guided Exercise

Reading an Event Study: Minimum Wage Increases and Teen Employment

A researcher estimates an event study of state-level minimum wage increases on teen employment rates. She plots coefficients for 4 periods before and 4 periods after the minimum wage change, normalizing the period just before the change (k = -1) to zero. The pre-period coefficients (k = -4 through k = -2) are all close to zero and not statistically significant. Post-period coefficients decline steadily from -0.01 at k = +1 to -0.04 at k = +4.

Why is the k = -1 period normalized to zero?

What does a flat pre-trend (coefficients near zero for k = -4 to k = -2) suggest?

The post-period coefficients grow in magnitude over time (-0.01 to -0.04). What does this tell you about the employment effect?

If the pre-period confidence intervals were very wide (e.g., ±0.30), why would a 'flat pre-trend' be uninformative?

Error Detective

Read the analysis below carefully and identify the errors.

A researcher studies the effect of state-level minimum wage increases on employment. They estimate an event study using TWFE with the specification: `Y_it = alpha_i + lambda_t + sum(beta_k * D_it^k) + epsilon_it`. Treatment timing is staggered across states from 2010 to 2018. The pre-treatment coefficients are all close to zero and not statistically significant. The researcher concludes: 'The flat pre-trends confirm parallel trends, and the average post-treatment effect of -0.03 represents the causal effect of minimum wage increases on employment.'

Select all errors you can find:

Referee Exercise

Read the paper summary below and write a brief referee critique (2-3 sentences) of the identification strategy.

Paper Summary

The authors study the effect of CEO succession on cumulative abnormal stock returns using an event-study design around 450 CEO transitions at S&P 500 firms from 2000-2020. They estimate abnormal returns using a market model with a 250-day estimation window and a [-10, +10] day event window around the announcement.

Key Table

Event WindowCARt-stat
[-10, -2]0.3%0.82
[-1, 0]1.8%3.41
[+1, +5]0.6%1.12
[+6, +10]-0.4%-0.73

Authors' Identification Claim

The near-zero pre-event CAR confirms that markets did not anticipate the succession, and the significant CAR at the announcement date identifies the causal effect of CEO change on firm value.


ISwap-In: When to Use Something Else

  • Canonical DiD: When a single treatment-effect estimate is sufficient and the time profile of effects is not of primary interest.
  • Staggered DiD estimators: When the goal is a single aggregate Average Treatment Effect on the Treated (ATT) across cohorts — modern estimators (Callaway & Sant'Anna, Sun & Abraham) aggregate event-study coefficients into a robust summary.
  • Synthetic control: When there is a single treated unit and constructing a counterfactual from donor units is more credible than assuming parallel trends.
  • Interrupted time series (ITS): When there is no control group and the design relies entirely on modeling the pre-treatment trend to forecast the counterfactual.

JReviewer Checklist

Critical Reading Checklist

0 of 8 items checked0%

Paper Library

Foundational (8)

Borusyak, K., Jaravel, X., & Spiess, J. (2024). Revisiting Event-Study Designs: Robust and Efficient Estimation.

Review of Economic StudiesDOI: 10.1093/restud/rdae007

Borusyak, Jaravel, and Spiess propose an imputation estimator for staggered DID that first estimates unit and time fixed effects from untreated observations, then imputes the counterfactual outcomes. This approach is efficient, flexible, and avoids the negative weighting problem of TWFE.

Brown, S. J., & Warner, J. B. (1985). Using Daily Stock Returns: The Case of Event Studies.

Journal of Financial EconomicsDOI: 10.1016/0304-405X(85)90042-X

Brown and Warner extend the event study framework from monthly to daily stock returns and examine the statistical properties of various test statistics. Their simulations show that simple methods perform well in most settings, providing practical reassurance for applied researchers.

Callaway, B., & Sant'Anna, P. H. C. (2021). Difference-in-Differences with Multiple Time Periods.

Journal of EconometricsDOI: 10.1016/j.jeconom.2020.12.001

Callaway and Sant'Anna propose group-time average treatment effects (ATT(g,t)) that avoid the problematic comparisons in TWFE. Their framework allows for heterogeneous treatment effects across groups and time and provides aggregation schemes for summary parameters.

Fama, E. F., Fisher, L., Jensen, M. C., & Roll, R. (1969). The Adjustment of Stock Prices to New Information.

International Economic ReviewDOI: 10.2307/2525569

Fama, Fisher, Jensen, and Roll establish the modern event study methodology by studying how stock prices adjust to stock splits. They develop the framework of measuring abnormal returns around corporate events using a market model to construct the counterfactual return. This methodology has become the standard tool for studying how information events affect asset prices and is used in thousands of subsequent studies across finance and strategy.

Freyaldenhoven, S., Hansen, C., & Shapiro, J. M. (2019). Pre-Event Trends in the Panel Event-Study Design.

American Economic ReviewDOI: 10.1257/aer.20180609

Freyaldenhoven, Hansen, and Shapiro study panel event-study designs in which unobserved confounds can generate pre-event trends. They show how causal effects can still be identified by exploiting covariates related to the policy only through the confounds, yielding a 2SLS estimator that remains valid even when endogeneity induces pre-trends.

Rambachan, A., & Roth, J. (2023). A More Credible Approach to Parallel Trends.

Review of Economic StudiesDOI: 10.1093/restud/rdad018

Rambachan and Roth develop a sensitivity analysis framework for assessing the robustness of event-study and difference-in-differences estimates to violations of the parallel trends assumption. Their approach constructs honest confidence intervals under restrictions on how pre-trends can extrapolate into the post-treatment period, providing a disciplined alternative to informal pre-trend tests.

Roth, J. (2022). Pretest with Caution: Event-Study Estimates after Testing for Parallel Trends.

American Economic Review: InsightsDOI: 10.1257/aeri.20210236

Roth shows that the common practice of testing for parallel pre-trends and proceeding conditional on 'passing' can lead to distorted inference. He proposes honest confidence intervals that account for pre-testing, fundamentally changing how researchers should think about event study pre-trends in DiD designs.

Sun, L., & Abraham, S. (2021). Estimating Dynamic Treatment Effects in Event Studies with Heterogeneous Treatment Effects.

Journal of EconometricsDOI: 10.1016/j.jeconom.2020.09.006

Sun and Abraham show that conventional event-study regression coefficients are contaminated by treatment effect heterogeneity across cohorts and propose an interaction-weighted estimator that recovers clean dynamic treatment effects. This paper is the key reference for event-study plots in staggered settings.

Application (2)

Capron, L., & Pistre, N. (2002). When Do Acquirers Earn Abnormal Returns?.

Strategic Management JournalDOI: 10.1002/smj.262

Capron and Pistre use event study methodology to examine when acquiring firms earn positive abnormal returns from mergers and acquisitions. They find that acquirers earn positive returns only when they are the primary source of value creation, contributing to the M&A strategy literature.

Dranove, D., & Olsen, C. (1994). The Economic Side Effects of Dangerous Drug Announcements.

Journal of Law and EconomicsDOI: 10.1086/467316

Dranove and Olsen use event studies to measure the stock market impact of FDA drug safety announcements on pharmaceutical firms. This application demonstrates how event studies can quantify the financial consequences of regulatory actions in health care and management contexts.

Survey (5)

Angrist, J. D., & Pischke, J.-S. (2009). Mostly Harmless Econometrics: An Empiricist's Companion.

Princeton University PressDOI: 10.1515/9781400829828

Angrist and Pischke write one of the most influential modern textbooks on applied econometrics, organizing the field around a design-based approach to causal inference. The book provides essential treatments of instrumental variables, difference-in-differences, and regression discontinuity, each grounded in the potential outcomes framework. It remains the standard reference for graduate students learning to evaluate and implement identification strategies.

de Chaisemartin, C., & D'Haultfoeuille, X. (2023). Two-Way Fixed Effects and Differences-in-Differences with Heterogeneous Treatment Effects: A Survey.

Econometrics JournalDOI: 10.1093/ectj/utac017

De Chaisemartin and D'Haultfoeuille provide a comprehensive survey of the recent literature on problems with two-way fixed effects estimators under heterogeneous treatment effects. They cover the key diagnostic tests (including the Goodman-Bacon decomposition), alternative estimators that are robust to heterogeneity, and practical guidance for choosing among them. The survey is essential reading for applied researchers working with event-study and difference-in-differences designs who need to understand when standard TWFE is and is not appropriate.

Kothari, S. P., & Warner, J. B. (2007). Econometrics of Event Studies.

Handbook of Empirical Corporate FinanceDOI: 10.1016/B978-0-444-53265-7.50015-9

Kothari and Warner provide an updated survey of event study methods, covering long-horizon event studies, cross-sectional regression approaches, and the econometric challenges that arise with overlapping events and event-induced variance changes. The survey documents how the basic FFJR framework is extended and refined over four decades. It is an essential reference for researchers designing event studies who need to understand the full menu of methodological choices and their trade-offs.

MacKinlay, A. C. (1997). Event Studies in Economics and Finance.

Journal of Economic Literature

MacKinlay provides a comprehensive methodological survey of event studies, covering the statistical framework, estimation windows, abnormal return calculations, and testing procedures. This paper remains the standard reference for researchers designing and implementing event studies.

McWilliams, A., & Siegel, D. (1997). Event Studies in Management Research: Theoretical and Empirical Issues.

Academy of Management JournalDOI: 10.2307/257056

McWilliams and Siegel provide a critical assessment of event study methodology as applied in management research, identifying common theoretical and design pitfalls including confounding events, improper event window selection, and thin trading. The paper outlines procedures for appropriate use of event studies and serves as a widely cited methodological guide for strategy and management researchers conducting event studies.

Tags

design-basedpaneldynamic-effects