Chapter 15: Causal Machine Learning

Interference, Spillovers, and SUTVA Violations in Financial Markets

In markets, one unit's treatment rarely stays politely confined to that unit.

Interference, Spillovers, and SUTVA Violations in Financial Markets

In markets, one unit's treatment rarely stays politely confined to that unit.

The Intuition

Many causal methods lean on a quiet assumption: one unit's treatment should not change another unit's outcome. In the potential-outcomes language, the outcome for unit i depends only on i's own treatment:

$$ Y_i = Y_i(T_i), $$

not on the whole treatment vector $(T_1,\dots,T_n)$.

This is part of SUTVA, the Stable Unit Treatment Value Assumption. In many policy or medical settings, it can be a reasonable starting point. In financial markets, it is often false.

Why? Because markets are equilibrium systems:

  • assets share macro drivers
  • trades move prices and funding conditions
  • information leaks before the formal event
  • portfolio rebalancing creates feedback into exposures and prices

That finance-specific fragility matters before you even get to estimator choice. A design that looks clean in a cross-sectional table may still be contaminated by shared exposure, benchmarked rebalancing, or pre-positioning underneath.

Why Finance Is Especially Fragile

Causal methods often assume units are separate enough that local treatment can be interpreted locally. Financial markets do the opposite:

  • common discount-rate shocks hit many assets together
  • sector and supply-chain links propagate firm-specific news
  • funding and liquidity conditions spread through correlated books
  • benchmarked investors create mechanically linked rebalancing

This is why Chapter 15 is more convincing when it talks about finance-specific failure modes than when it simply lists methods. The methods are standard. The market structure is what breaks them.

What SUTVA Violation Looks Like

The clean version of SUTVA says:

$$ Y_i(T_i, T_{-i}) = Y_i(T_i), $$

where $T_{-i}$ denotes everyone else's treatment. A violation means the outcome for unit i depends on others' treatment too.

In markets, this happens in several ways.

Spillovers

A macro announcement affects not only the treated bond ETF but also other rate-sensitive assets used as controls.

Interference through rebalancing

If many portfolios respond to the same factor signal, one asset's "treatment" changes demand in related assets and feeds back into returns.

Anticipation

If traders position before the official event time, the treatment begins before the timestamp used in the design.

Equilibrium feedback

Factor exposures affect prices, which then change future factor exposures through reweighting, constraints, and benchmark effects.

These are not edge cases. They are routine.

A Worked Example

Suppose you want to measure the impact of an FOMC surprise on a Treasury ETF and you choose credit ETFs and REITs as controls.

The naive story is:

  • treated asset: Treasury ETF
  • controls: other assets not directly targeted
  • event: Fed announcement

But in reality the shock moves rates, credit spreads, discount rates, and risk appetite across the whole system. The controls are not untreated. They are differently treated.

Now imagine the event study finds a clear gap between the treated ETF and the BSTS counterfactual. That gap may still be useful descriptively, but the identifying assumption is weaker than it looks because the counterfactual is built from controls hit by the same shock.

The same problem appears in factor research. A momentum tilt may seem to cause higher returns, but if the portfolio also acquires incidental tech concentration and market beta, the treatment is not a clean isolated intervention. It is embedded in a network of shared exposures.

Anticipation Is a Hidden Spillover

Anticipation is worth separating because it often hides inside otherwise careful designs.

If traders act before an earnings release or macro announcement, then the treatment does not start at the timestamp in the dataset. The design is effectively measuring an event after part of its effect has already been absorbed by prices.

Options-implied volatility often moves before an earnings release, and Fed Funds futures reprice before the formal FOMC statement when the market infers the likely outcome. In both cases, using the official timestamp as t=0 understates how much of the treatment was already in the market.

This matters because:

  • placebo windows can look active before the event
  • pre-trend checks can fail even if the mechanism is real
  • a "no effect at announcement time" result may reflect early incorporation, not no effect

In finance, event timing is therefore part of identification, not a clerical detail.

What Mitigation Can Really Do

There is no universal fix for interference in markets. Mitigation is partial, not curative.

Still, some design choices help:

  • use surprise-based treatments instead of scheduled-event indicators when possible
  • prefer relative outcomes over raw levels when the mechanism is comparative
  • use narrower windows when anticipation and spillover grow quickly with time
  • test robustness across control universes and sample definitions
  • use negative-control outcomes to check whether the same shared driver is leaking elsewhere

The important thing is intellectual honesty. If spillovers are structurally unavoidable, the claim should be framed as fragile or local, not as a clean unit-level treatment effect.

A Practical Heuristic

Ask three questions before trusting a finance causal design:

  1. Could the event or treatment plausibly move the controls too?
  2. Could market participants have acted before the official treatment timestamp?
  3. Could the treatment change prices or exposures in ways that feed back into other units?

If the answer to any is yes, SUTVA is under pressure. The design may still be useful, but the causal language should become more cautious.

In Practice

Keep the claims proportional to the design:

  • treat spillover checks and alternative universe definitions as first-order diagnostics, not appendix material
  • be willing to downgrade the claim from "causal effect" to "suggestive directional evidence" when interference is structurally hard to rule out
  • align event windows to the mechanism, not to convenient calendar bins

Common Mistakes

  • Talking as if controls are untreated simply because they were not the main asset of interest.
  • Ignoring anticipation because the official event timestamp is easy to observe.
  • Treating equilibrium feedback as a secondary implementation detail.
  • Presenting a fragile finance design with the same confidence as a clean randomized setup.
  • Assuming robustness across one universe definition is enough.

Connections

This primer supports Chapter 15's identification pitfalls and BSTS spillover critique. It connects directly to placebo and negative-control diagnostics, event-study design, and the broader question of when a predictive relation can plausibly be interpreted causally.

Register to Read

Sign up for a free account to access all 61 primer articles.

Create Free Account

Already have an account? Sign in