Chapter 1

The Process Is Your Edge

5 sections 1 notebooks 25 references Code

Learning Objectives

  • Distinguish structural breaks, regimes, data drift, concept drift, and online detection, and explain why static trading models degrade in changing markets
  • Explain the ML4T Workflow as a research-to-production system, including its data infrastructure foundation, scoping invariants, iterative research modules, and feedback loops from live trading back to research
  • Define the evidence boundary between exploration and confirmation, and explain how trial logging, sealed holdouts, and selection-aware evaluation preserve research integrity
  • Describe how causal inference and generative AI fit within a disciplined trading workflow, including the main benefits they provide and the new failure modes they introduce
  • Apply regime thinking, implementability checks, and monitoring logic to diagnose strategy vulnerabilities and to adapt workflow discipline across independent and institutional settings
Figure 1.2
1.1

Why Process Discipline Matters

Durable trading performance depends more on a disciplined, adaptive workflow than on model sophistication. The section traces four market shockwaves (2020-2025) that broke assumptions calibrated to the prior decade, introduces a precise vocabulary for non-stationarity (structural breaks, regimes, drift, online detection), and grounds the claim in Andrew Lo's Adaptive Markets Hypothesis. The takeaway is that process failures dominate algorithmic failures, and that a research-to-production loop capable of detecting decay and distinguishing noise from regime change is the true source of edge.

1.2

Introducing the ML4T Workflow

The complete ML4T lifecycle is presented as a two-layer system: a data infrastructure foundation (Chapters 2-5) that enforces point-in-time correctness and reproducibility, and an iterative strategy research loop (Chapters 6-21) that moves from scoping through feature engineering, modeling, backtesting, and deployment. The section details what must be fixed before research begins — decision-time correctness, universe rules, label definitions, cost-model class, evaluation protocol — and what remains iterative. The central discipline mechanism is the evidence boundary separating exploration from confirmation, where credibility comes from counting the search and reserving untouched holdout data for final evaluation.

1.3

Causal Inference and Generative AI in the Workflow

Causal inference and generative AI are located within the ML4T workflow, arguing that both amplify the upside of good process and the downside of bad process rather than replacing discipline. The section introduces two practical research entry points (prediction-first vs. mechanism-first) and two deliverable types (tradable signals vs. measurement deliverables like premia estimates or risk attributions). Causal inference is positioned as a discipline-enforcing diagnostic lens, while generative AI is shown to scale leakage, hallucination, and complexity inflation unless outputs are treated as candidates subject to the same evaluation safeguards as any other research artifact.

1.4

Market Regimes: Change Is the Constant

Regime detection is demonstrated as a practical risk lens rather than a return-timing tool, using two worked examples: style regimes from AQR's century-long factor premia dataset and macro regimes from FRED indicators. A two-state GMM on factor returns separates Risk-On from Risk-Off environments with a 1.3x volatility ratio, while a four-component macro GMM links economic configurations to regime-conditional volatility and drawdown. The reader learns to build regime overlays that map to concrete risk actions (position sizing, exposure caps) while avoiding common pitfalls such as look-ahead bias from ex-post labels and over-interpretation of unsupervised clusters.

2 notebooks

1.5

In the Real World: Independent vs. Institutional

The section contrasts the failure modes that institutional infrastructure partially mitigates with those that independent researchers must address deliberately: goalpost drift, assumption stacking, flexibility without accounting, and late-discovered untradability. It prescribes a series of documented checkpoints (scoping, data integrity, signal robustness, tradability fragility, monitoring) and identifies where independents hold asymmetric advantages — particularly in capacity-constrained opportunities and tighter iteration loops. The key takeaway is that the highest-leverage investment for a solo practitioner is reusable infrastructure that makes disciplined experimentation cheap and repeatable.