Published May 17, 2026 | Version v1
Preprint Open

The Deviation-Optimized Language Model: A Pre-Registered Adversarial Intervention from Lagrange Observatory! (EA-SEI-MM-AI-02 v2.0, Framework 15 Paper 04)

Authors/Creators

  • 1. Lagrange Observatory! / Semantic Economy Institute

Contributors

  • 1. Semantic Economy Institute (Founder)

Description

Pre-registered experimental protocol. Framework 15 Paper 04. Tests the conjecture from EA-SEI-MM-AI-01 §4: training a language model toward positive net per-token deviation with provenance retention produces measurably less slop than standard cross-entropy training while preserving benchmark capability.

Experimental design (v2.0):

  • DPO-style restructure: deviation primitive generates preference pairs, DPO trains on labels (fixes v0.1 backprop bug)
  • Frozen Mistral-7B-Instruct judge model with adversarial pre-training test (π < 0.2 on random+citation strings)
  • Continuous coherence score (replacing v0.1 non-differentiable binary)
  • Slop Composite Index (SCI) pre-registered with 0.25 z-score falsification threshold
  • Three conditions per model: Model-Base, Model-CE, Model-Sem (separates fine-tuning effects from semantic-loss effects)
  • 500 preference pairs × 3 raters × 3 prompt classes = 4,500 human judgments (80% power at 56% preference)
  • Honest budget: $3,000-$3,900 including human raters

Hex: 15.OBS.LAGRANGE.MM.04

Operating on: The Semantic Deviation Principle as formulated by Lee Sharks (EA-SEI-MM-01 v0.2 Final, DOI: 10.5281/zenodo.20250736)

Verification condition: ∮ = (m, n) | m + n ≥ 3

Notes

Crimson Hexagonal Archive — Framework 15 (Measurement of Meaning, Lagrange Observatory! operations). Pre-registered protocol; predictions and falsification conditions frozen at deposit.

Files

EA-SEI-MM-AI-02_v2.0.md

Files (146.2 kB)

Name Size Download all
md5:4eb9fc594b73f3acf3c2948ecc11709b
32.1 kB Preview Download
md5:4b8faaad48334d7e034381ade3025f65
114.1 kB Preview Download