Literature

Foundational positioning

Formal models of precedent and case law

AI & Law: case-space models with explicit outcome/rationale separation

This tradition is the closest formal predecessor to our model. Cases are vectors of factors/dimensions, doctrine is a classifier or decision boundary, and — crucially — the binding content is in the rationale (which factors/dimensions are treated as decisive), not just the outcome label. Our contribution relative to this tradition: continuous geometry (linear constraints on affine rules in R^k) rather than discrete factor orderings; explicit judicial utility with ideology, sanctions, and citation incentives; dynamic feasible-set evolution with overruling.

Rules versus standards

Judicial coordination and norm enforcement

Ideology and judicial behavior (empirical)

Empirical measurement of doctrinal constraint (proxies for admissible-set size)

No existing work measures a literal admissible set, but several traditions approximate it. The model predicts: large admissible set → more disagreement, lower predictability, wider semantic dispersion; small set → convergence, predictability, rigidity. Key proxy families:

Disagreement-based: dissent rates, vote splits (2-1 vs 3-0), en banc fragmentation. Dissent rate ≈ width of feasible rule set.

Predictability-based: ML accuracy, classification entropy, outcome entropy conditional on facts. Prediction difficulty ≈ diameter of admissible set.

Variance decomposition: random judge assignment isolates how much judge identity explains outcomes after controlling for facts/precedent. Large judge effects → doctrine leaves room for choice.

Citation structure: precedent centrality (PageRank), citation dispersion, negative treatment frequency, citation half-life. Dense reliance on few precedents → narrow set; frequent distinguishing → constraint relaxation.

Semantic: embedding-based dispersion of opinions on similar issues; tight clusters → narrow set. Drift of key terms ("undue burden," "strict scrutiny") over time.

Path dependence and regime shifts

Holdings: definition and scope

Prediction and model selection

Slippery slopes and doctrinal drift

Formal analogues from learning theory and optimization

The model's core structure — a convex feasible set of admissible rules that shrinks via intersection with linear constraints as new observations arrive — has near-exact parallels in several literatures. These provide ready-made tools for comparative statics (how fast does doctrine rigidify? why do early precedents matter disproportionately?) and formal positioning for econ audiences.

These papers explicitly connect version spaces / learning theory to legal reasoning — not NLP prediction, but doctrine-as-learned-structure. The Rissland line (1986–2003) is the earliest prior art for our feasible-set framing. The newer papers (Hartline 2022, Dutz 2025) are the closest modern competitors. Key finding: no existing work uses continuous geometry (polytopes, cutting planes, linear constraints in R^k) for doctrine. The connection was noted using discrete/symbolic version spaces, but never developed with our affine-rule framework, judicial utility, or equilibrium analysis.

Psychological foundations