ArXiv TLDR

Stacked Triple Differences

🐦 Tweet
2604.22982

Meng Hsuan Hsieh

econ.EM

TLDR

Stacked Triple Differences (DDD) resolves issues in conventional DDD under staggered adoption, offering a transparent, regression-based approach for causal inference.

Key contributions

  • Resolves forbidden-comparison and interpretation issues in conventional DDD under staggered adoption.
  • Extends stacked difference-in-differences to DDD using self-contained four-cell event window stacks.
  • Identifies a positive, cell-size-weighted average of stack-level conditional average treatment effects.
  • Provides regression-based transparency, pairwise parallel trends, and direct control over aggregation weights.

Why it matters

This paper introduces Stacked DDD, a novel method addressing critical issues in conventional triple differences under staggered adoption. It offers a transparent, regression-based alternative to existing GMM and imputation frameworks, providing clearer causal estimands. Empirical illustrations show it can lead to substantially different quantitative conclusions, improving causal inference.

Original Abstract

Triple differences (DDD) is a workhorse quasi-experimental design in applied economics. But, under staggered adoption, its conventional three-way fixed-effects (3WFE) implementation inherits the forbidden-comparison and interpretation issues now well understood in the difference-in-differences literature. To resolve these issues, I introduce stacked DDD. I extend the stacked difference-in-differences approach to the DDD setting by creating self-contained stacks, each consisting of four cells over an event window: treated and clean comparison cohorts, each with treatment-eligible and treatment-ineligible units. Appending these stacks yields a unified dataset for estimating treatment effects without making forbidden comparisons. I prove that, at each post-treatment event-time, a linear regression with fully saturated fixed-effects applied to the stacked dataset identifies a strictly positive, cell-size-weighted average of stack-level conditional average treatment effects, with stack weights proportional to stack-level cell sizes. Building on this characterization, I outline alternative weighting schemes that recover distinct, transparent causal estimands with clear interpretations. Stacked DDD complements recent GMM and imputation-based frameworks by trading efficiency for regression-based transparency, pairwise (rather than global) parallel trends, and direct control over aggregation weights. I provide two empirical illustrations where stacked DDD yields substantially different quantitative conclusions compared to existing procedures.

📬 Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.