MethodAtlas
Design-BasedModern

Event Studies (Dynamic Treatment Effects)

Visualize how treatment effects evolve over time — and test whether pre-trends support the parallel trends assumption.

Quick Reference

When to Use
When you want to visualize the time path of treatment effects and test for pre-existing differential trends. An event study is the dynamic generalization of DiD.
Key Assumption
Parallel trends in the pre-period (no differential trends before treatment) and no anticipation effects. Pre-treatment coefficients near zero support but do not prove this assumption.
Common Mistake
Interpreting non-significant pre-trend coefficients as proof that parallel trends holds (absence of evidence is not evidence of absence). Also, using standard TWFE event studies with staggered treatment timing without checking for heterogeneity bias.
Estimated Time
2.5 hours

One-Line Implementation

Stata: reghdfe y ib(-1).event_time, absorb(unit_id year) vce(cluster state)
R: feols(y ~ i(event_time, ref = -1) | unit_id + year, data = df, vcov = ~state)
Python: pf.feols('y ~ i(event_time, ref=-1) | unit + year', data=df, vcov={'CRV1': 'state'})

Download Full Analysis Code

Complete scripts with diagnostics, robustness checks, and result export.

Motivating Example

A publicly traded firm announces that its longtime CEO is stepping down and a new CEO will take over. You want to know: what is the causal effect of CEO succession on stock prices?

You cannot just compare stock prices before and after the announcement — maybe the whole market was going up, or maybe the firm was already in decline. A simple before-after comparison confounds the succession effect with everything else happening at the same time.

What you need is a picture. You want to see what happened to the stock price in the weeks before the announcement (to check whether something was already going on) and in the weeks after (to see the actual effect and how it evolves). Did the stock jump immediately? Did it drift? Did the effect fade or grow?

That time-profile picture is an event study.

A. Overview: What Is an Event Study?

An event study is a regression framework that estimates separate treatment effects for each time period relative to the event. Instead of getting one number ("the treatment effect is 3.2 percentage points"), you get a time profile: the effect at the moment of treatment, one period after, two periods after, and so on — plus coefficients for the periods before treatment that serve as a diagnostic check.

The core idea is elegantly simple. You define event time as the number of periods relative to when the treatment happens. If a state adopts a policy in 2010, then for that state, event time -2 is 2008, event time 0 is 2010, and event time +3 is 2013. You then estimate a coefficient for each event-time indicator.

In finance, event studies have been used since at least Fama et al. (1969).

In the broader social sciences, event-study designs have become the standard way to visualize treatment dynamics in difference-in-differences settings. In many applied fields, an event-study plot is now a standard expectation in DiD papers.

The Static DiD Problem

Recall the canonical 2x2 DiD specification:

Yit=αi+λt+δDit+εitY_{it} = \alpha_i + \lambda_t + \delta \cdot D_{it} + \varepsilon_{it}

This specification gives you a single coefficient δ\delta that averages the treatment effect across all post-treatment periods. But what if the effect takes three years to materialize? What if it fades after one year? What if it was already trending upward before treatment? The static model hides all of this dynamic information.

Common Confusions

"Which period should I omit as the reference?" The standard practice is to omit the period immediately before treatment (event time -1). This normalization ensures all coefficients are measured relative to the last pre-treatment period. Some researchers omit -2 instead, which is fine — just be consistent and transparent.

"Should I bin the endpoints?" If you have observations far from the event (e.g., event time -10 or +15), you typically bin them: group all observations at or below some threshold into a single indicator (e.g., event time 6\leq -6) and similarly for the right tail. This binning avoids estimating noisy coefficients from small samples at the extremes. But be careful — the binned coefficient averages effects across all those periods, which can be misleading if there are trends in the tails.

"Is an event study different from DiD?" An event study is a DiD. It is the dynamic generalization. The static DiD is a special case where you constrain all post-treatment effects to be the same and all pre-treatment effects to be zero.

B. Identification

The event-study specification is:

Yit=αi+λt+k1βk1[Kit=k]+εitY_{it} = \alpha_i + \lambda_t + \sum_{k \neq -1} \beta_k \cdot \mathbf{1}[K_{it} = k] + \varepsilon_{it}

where αi\alpha_i and λt\lambda_t are unit and time fixed effects, KitK_{it} is the event time for unit ii at calendar time tt, and the sum runs over all event-time indicators except the reference period k=1k = -1.

Identifying assumptions:

  1. Parallel trends: In the absence of treatment, treated and control units would have followed the same trajectory. Formally, the counterfactual trend in outcomes absent treatment is the same for treated and control groups: E[Yit(0)Yi,t1(0)Di=1]=E[Yit(0)Yi,t1(0)Di=0]E[Y_{it}(0) - Y_{i,t-1}(0) \mid D_i = 1] = E[Y_{it}(0) - Y_{i,t-1}(0) \mid D_i = 0].

  2. No anticipation: The treatment has no effect before it is actually implemented. This condition means βk=0\beta_k = 0 for all k<0k < 0.

The pre-treatment coefficients {βk}k<1\{\beta_k\}_{k < -1} are your diagnostic: if they are jointly close to zero, this pattern is consistent with both parallel trends and no anticipation. If they show a clear trend, either parallel trends or no anticipation (or both) may be violated.

C. Visual Intuition

The classic event-study plot has event time on the x-axis and coefficient estimates (with 95% confidence intervals) on the y-axis. The reference period (typically k=1k = -1) is normalized to zero.

A good event study has the following pattern: pre-treatment coefficients hover around zero with confidence intervals that include zero, then there is a clear jump at k=0k = 0 that persists (or grows, or fades) in the post-treatment periods.

A problematic event study has the following pattern: pre-treatment coefficients trend upward (or downward), suggesting that treated units were already diverging from control units before the treatment. If you see such a pattern, your parallel trends assumption is in trouble.

Interactive Simulation

Event Study Anatomy

Adjust the true treatment effect, the presence of pre-trends, and the noise level. Watch how the event-study plot changes. With pre-trends, can you still tell what the treatment effect is?

-2.82-0.831.173.175.17Test Statistic (Difference in Means)051101152CountObserved: 4.500p = 0.001Observed StatisticTail (p-value region)Permutation Null
010
-22
0.55
50500
Interactive Simulation

Why Event Study Design?

Panel DiD with 200 units (100 treated), 10 periods, treatment at t = 5. Anticipation = 0.0, dynamic growth = 0.0.

Treatment-1.33-0.350.621.602.573.54-5-4-3-2-101234Relative Time to TreatmentCoefficient
Static DiDEvent StudyTrue Effect

Estimation Results

Estimatorβ̂SE95% CIBias
Static DiDclosest2.0730.197[1.69, 2.46]+0.073
Event Study (avg post)2.0990.197[1.71, 2.48]+0.099
True β2.000
0.0

0 = no anticipation; 1 = strong anticipation (violates parallel trends)

0.0

0 = constant effect; positive = growing; negative = fading

2.0

Base treatment effect at time 0

200

Total units (half treated). Larger N → less sampling noise.

Why the difference?

Pre-trend coefficients are all statistically insignificant (close to zero), supporting the parallel trends assumption. This is the key diagnostic that event studies provide over static DiD. With constant treatment effects, the static DiD (β̂ = 2.073) closely matches the true ATT (2.000). Both approaches recover the treatment effect well; the event study additionally validates the parallel-trends assumption. Static DiD deviation from true ATT: 0.073. The event study reveals the full dynamic path of treatment effects.

D. Mathematical Derivation

Don't worry about the notation yet — here's what this means in words: Each event-study coefficient estimates the average difference between treated and control units at a specific time relative to treatment, after removing unit-specific and time-specific components.

Start with the potential outcomes framework. For each unit ii at time tt, the observed outcome is:

Yit=Yit(0)+kβkDitkY_{it} = Y_{it}(0) + \sum_{k} \beta_k \cdot D_{it}^k

where Ditk=1[Kit=k,k0]D_{it}^k = \mathbf{1}[K_{it} = k, k \geq 0] indicates being kk periods after treatment. Under parallel trends:

E[Yit(0)i,t]=αi+λtE[Y_{it}(0) | i, t] = \alpha_i + \lambda_t

Substituting and adding pre-treatment indicators as diagnostic terms:

Yit=αi+λt+k1βk1[Kit=k]+εitY_{it} = \alpha_i + \lambda_t + \sum_{k \neq -1} \beta_k \cdot \mathbf{1}[K_{it} = k] + \varepsilon_{it}

Under correct specification, the OLS estimator for βk\beta_k (for k0k \geq 0) identifies the average treatment effect at event time kk:

β^kpE[Yit(1)Yit(0)Kit=k]\hat{\beta}_k \xrightarrow{p} E[Y_{it}(1) - Y_{it}(0) | K_{it} = k]

For k<1k < -1, the coefficients βk\beta_k should be zero under the null of parallel trends and no anticipation. A joint F-test of all pre-treatment coefficients provides a formal test.

Important caveat ((Sun & Abraham, 2021)): When treatment timing is staggered, the OLS estimates of βk\beta_k from the specification above can be contaminated. Already-treated units serve as controls for later-treated units, leading to negative weights on some group-time treatment effects. The interaction-weighted estimator of Sun and Abraham (2021) fixes this by estimating cohort-specific effects and then aggregating properly.

E. Implementation

library(fixest)

# Standard event study using fixest
# i(event_time, ref = -1) creates event-time dummies with k=-1 as reference
# Unit and year fixed effects are absorbed via | unit_id + year
# Standard errors are clustered at the state level
est <- feols(y ~ i(event_time, ref = -1) | unit_id + year,
           data = df, vcov = ~state)

# Plot event-study coefficients with 95% confidence intervals
iplot(est, main = "Event Study: Treatment Effect Over Time")

# Sun & Abraham (2021) estimator for staggered treatment timing
# sunab(cohort_var, time_var) avoids contamination from already-treated units
# treatment_year = year each unit was first treated (Inf if never treated)
est_sa <- feols(y ~ sunab(treatment_year, year) | unit_id + year,
              data = df, vcov = ~state)
iplot(est_sa)
Requiresfixest

F. Diagnostics

  1. Joint F-test of pre-treatment coefficients. Test whether β2,β3,\beta_{-2}, \beta_{-3}, \ldots are jointly zero. This joint test is more informative than eyeballing individual coefficients. In Stata: testparm on the pre-treatment indicators. In R with fixest: wald(est, "event_time::-").

  2. Visual inspection of pre-trends. Even if the F-test does not reject, look at the pattern. A clear monotonic trend that approaches zero at k=1k = -1 is concerning even if individual coefficients are not significant.

  3. Confidence interval widths. If your pre-treatment confidence intervals are wide enough to include economically meaningful effects, your pre-trend test lacks power. You cannot distinguish "no pre-trend" from "pre-trend we cannot detect."

  4. Placebo treatment dates. Artificially reassign treatment to earlier periods and re-estimate. If you still see "effects," something is wrong.

  5. Sensitivity to endpoint binning. Re-estimate with different bin cutoffs. If your results change substantially, the specification is fragile.

Interpreting Your Results

Pre-treatment coefficients near zero: This finding supports (but does not prove) parallel trends. Report the joint F-test p-value alongside the visual.

Pre-treatment coefficients trending: This pattern is a red flag. Consider whether the trend is economically meaningful. If it is large relative to your post-treatment effects, your results are not credible without further adjustment (e.g., including unit-specific linear trends, though this introduces its own problems).

Post-treatment effects that grow over time: This dynamic is common and often substantively meaningful. A policy may take time to have its full effect. Report the dynamics, not just the average.

Post-treatment effects that shrink: This pattern could mean the effect is transitory, or that control units are "catching up." Think about the mechanism.

G. What Can Go Wrong

Assumption Failure Demo

Staggered Timing Contaminates Event-Study Coefficients

Researcher uses the Sun and Abraham (2021) interaction-weighted estimator for an event study of state health-insurance expansions adopted between 2010 and 2018, with a never-treated control group.

Post-treatment coefficients show a gradual increase in insurance coverage: +1.2 pp at k=0, +2.8 pp at k=1, +4.1 pp at k=2, consistent with phased enrollment dynamics.

Assumption Failure Demo

Underpowered Pre-Trend Test Creates False Confidence

Researcher estimates an event study of CEO turnovers on firm investment rates with 200 firms and 10 pre-treatment quarters. They report both the event-study plot and the power of the pre-trend test, noting that the 95% CI at each pre-period ranges from -1.5 to +1.5 percentage points.

The researcher acknowledges: 'Our pre-trend test cannot rule out differential pre-trends as large as 1.5 pp per quarter, which over 4 quarters could generate a cumulative bias of 6 pp — comparable to our estimated post-treatment effect of 4.2 pp. We therefore view the parallel trends assumption as plausible but not definitive.'

Assumption Failure Demo

Endpoint Binning Masks Problematic Dynamics

Researcher studies the effect of plant closings on local unemployment using event time from -8 to +8 years. They bin endpoints at k <= -6 and k >= 6, and separately report the un-binned specification as a robustness check.

The binned specification shows a clean pattern. The un-binned specification reveals that coefficients at k = -7 and k = -8 are trending upward, suggesting potential anticipation or pre-existing decline. The researcher investigates and finds that some closings were preceded by gradual layoffs.

H. Practice

Concept Check

In an event-study regression, the coefficient at event time k = -3 is 0.02 with a standard error of 0.15. The coefficient at event time k = +2 is 0.08 with a standard error of 0.03. What should you conclude?

Guided Exercise

Reading an Event Study: Minimum Wage Increases and Teen Employment

A researcher estimates an event study of state-level minimum wage increases on teen employment rates. She plots coefficients for 4 periods before and 4 periods after the minimum wage change, normalizing the period just before the change (k = -1) to zero. The pre-period coefficients (k = -4 through k = -2) are all close to zero and not statistically significant. Post-period coefficients decline steadily from -0.01 at k = +1 to -0.04 at k = +4.

Why is the k = -1 period normalized to zero?

What does a flat pre-trend (coefficients near zero for k = -4 to k = -2) suggest?

The post-period coefficients grow in magnitude over time (-0.01 to -0.04). What does this tell you about the employment effect?

If the pre-period confidence intervals were very wide (e.g., ±0.30), why would a 'flat pre-trend' be uninformative?

Error Detective

Read the analysis below carefully and identify the errors.

A researcher studies the effect of state-level minimum wage increases on employment. They estimate an event study using TWFE with the specification: Y_it = alpha_i + lambda_t + sum(beta_k * D_it^k) + epsilon_it. Treatment timing is staggered across states from 2010 to 2018. The pre-treatment coefficients are all close to zero and not statistically significant. The researcher concludes: 'The flat pre-trends confirm parallel trends, and the average post-treatment effect of -0.03 represents the causal effect of minimum wage increases on employment.'

Select all errors you can find:

Referee Exercise

Read the paper summary below and write a brief referee critique (2-3 sentences) of the identification strategy.

Paper Summary

The authors study the effect of CEO succession on cumulative abnormal stock returns using an event-study design around 450 CEO transitions at S&P 500 firms from 2000-2020. They estimate abnormal returns using a market model with a 250-day estimation window and a [-10, +10] day event window around the announcement.

Key Table

Event WindowCARt-stat
[-10, -2]0.3%0.82
[-1, 0]1.8%3.41
[+1, +5]0.6%1.12
[+6, +10]-0.4%0.73

Authors' Identification Claim

The near-zero pre-event CAR confirms that markets did not anticipate the succession, and the significant CAR at the announcement date identifies the causal effect of CEO change on firm value.

I. Swap-In: When to Use Something Else

  • Canonical DiD: When a single treatment-effect estimate is sufficient and the time profile of effects is not of primary interest.
  • Staggered DiD estimators: When the goal is a single aggregate ATT across cohorts — modern estimators (Callaway & Sant'Anna, Sun & Abraham) aggregate event-study coefficients into a robust summary.
  • Synthetic control: When there is a single treated unit and constructing a counterfactual from donor units is more credible than assuming parallel trends.
  • Interrupted time series (ITS): When there is no control group and the design relies entirely on modeling the pre-treatment trend to forecast the counterfactual.

J. Reviewer Checklist

Critical Reading Checklist

Paper Library

Foundational (9)

Fama, E. F., Fisher, L., Jensen, M. C., & Roll, R. (1969). The Adjustment of Stock Prices to New Information.

International Economic ReviewDOI: 10.2307/2525569

This paper is the origin of the modern event study methodology in finance. Fama, Fisher, Jensen, and Roll studied how stock prices adjust to stock splits and established the basic framework of measuring abnormal returns around corporate events that has been used in thousands of subsequent studies.

MacKinlay, A. C. (1997). Event Studies in Economics and Finance.

Journal of Economic Literature

MacKinlay provided a comprehensive methodological survey of event studies, covering the statistical framework, estimation windows, abnormal return calculations, and testing procedures. This paper remains the standard reference for researchers designing and implementing event studies.

Brown, S. J., & Warner, J. B. (1985). Using Daily Stock Returns: The Case of Event Studies.

Journal of Financial EconomicsDOI: 10.1016/0304-405X(85)90042-X

Brown and Warner extended the event study framework from monthly to daily stock returns and examined the statistical properties of various test statistics. Their simulations showed that simple methods perform well in most settings, providing practical reassurance for applied researchers.

Kothari, S. P., & Warner, J. B. (2007). Econometrics of Event Studies.

Handbook of Empirical Corporate FinanceDOI: 10.1016/B978-0-444-53265-7.50015-9

Kothari and Warner updated the survey of event study methods, covering long-horizon event studies, cross-sectional regression approaches, and the econometric challenges that arise with overlapping events and event-induced variance changes.

Borusyak, K., Jaravel, X., & Spiess, J. (2024). Revisiting Event-Study Designs: Robust and Efficient Estimation.

Review of Economic StudiesDOI: 10.1093/restud/rdae007

Borusyak, Jaravel, and Spiess proposed an imputation estimator for event studies that first estimates unit and time fixed effects from untreated observations, then imputes counterfactual outcomes for treated units. This approach is efficient, avoids negative weighting problems of two-way fixed effects, and is now a leading method for robust event-study estimation.

Rambachan, A., & Roth, J. (2023). A More Credible Approach to Parallel Trends.

Review of Economic StudiesDOI: 10.1093/restud/rdad018

Rambachan and Roth developed a sensitivity analysis framework for assessing the robustness of event-study and difference-in-differences estimates to violations of the parallel trends assumption. Their approach constructs honest confidence intervals under restrictions on how pre-trends can extrapolate into the post-treatment period, providing a disciplined alternative to informal pre-trend tests.

Freyaldenhoven, S., Hansen, C., & Shapiro, J. M. (2019). Pre-Event Trends in the Panel Event-Study Design.

American Economic ReviewDOI: 10.1257/aer.20180609

Freyaldenhoven, Hansen, and Shapiro developed diagnostic tools and an instrumental-variables-based estimator for panel event-study designs when pre-event trends may be present. Their framework helps researchers distinguish true anticipation effects from confounding trends, addressing a central challenge in event-study credibility.

Sun, L., & Abraham, S. (2021). Estimating Dynamic Treatment Effects in Event Studies with Heterogeneous Treatment Effects.

Journal of EconometricsDOI: 10.1016/j.jeconom.2020.09.006

The foundational paper for understanding why standard event-study regressions can break down under staggered treatment timing with heterogeneous effects. When already-treated units serve as controls for later-treated units, the TWFE event-study coefficients can be contaminated. Sun and Abraham propose the interaction-weighted estimator as a solution that properly aggregates cohort-specific effects.

Roth, J. (2022). Pretest with Caution: Event-Study Estimates after Testing for Parallel Trends.

American Economic Review: InsightsDOI: 10.1257/aeri.20210236

Demonstrates that conditioning on passing a pre-trend test can distort post-treatment estimates. When researchers run a pre-trend test and proceed only if it passes, the resulting estimates are biased because the conditioning event is informative about the magnitude of any underlying trend. Proposes honest confidence intervals that account for pre-testing.

Application (4)

McWilliams, A., & Siegel, D. (1997). Event Studies in Management Research: Theoretical and Empirical Issues.

Academy of Management JournalDOI: 10.2307/257056

McWilliams and Siegel introduced event study methods to the management research community, explaining the assumptions, methodology, and common pitfalls. This tutorial article led to widespread adoption of event studies in strategic management research.

Koh, P.-S., & Reeb, D. M. (2015). Missing R&D.

Journal of Accounting and EconomicsDOI: 10.1016/j.jacceco.2015.03.004

Koh and Reeb used event study methodology to examine stock market reactions to firms that report versus do not report R&D expenditures, showing that missing R&D data is not random and has implications for how investors value innovation. This paper illustrates event study methods applied to accounting and disclosure questions.

Dranove, D., & Olsen, C. (1994). The Economic Side Effects of Dangerous Drug Announcements.

Journal of Law and EconomicsDOI: 10.1086/467316

Dranove and Olsen used event studies to measure the stock market impact of FDA drug safety announcements on pharmaceutical firms. This application demonstrated how event studies can quantify the financial consequences of regulatory actions in health care and management contexts.

Capron, L., & Pistre, N. (2002). When Do Acquirers Earn Abnormal Returns?.

Strategic Management JournalDOI: 10.1002/smj.262

Capron and Pistre used event study methodology to examine when acquiring firms earn positive abnormal returns from mergers and acquisitions. They found that acquirers earn positive returns only when they are the primary source of value creation, contributing to the M&A strategy literature.

Survey (1)

de Chaisemartin, C., & D'Haultfoeuille, X. (2023). Two-Way Fixed Effects and Differences-in-Differences with Heterogeneous Treatment Effects: A Survey.

Econometrics JournalDOI: 10.1093/ectj/utac017

De Chaisemartin and D'Haultfoeuille provided a comprehensive survey of the recent literature on problems with two-way fixed effects estimators under heterogeneous treatment effects, covering the key diagnostic tests, alternative estimators, and practical guidance for applied researchers working with event-study and difference-in-differences designs.

Tags

design-basedpaneldynamic-effects