φ-Dynamics in Large Language Models: An Inductive Bias from Topology and the Theory of Diophantine Approximation
Authors/Creators
Description
Abstract
Large language models exhibit a persistent drop in output quality when faced with inputs that lie outside the training distribution - the so-called Out-of-Distribution (OOD) regime. The standard explanation is statistical: a mismatch between training and test distributions. We argue that there is also a geometric component: the Cartesian geometry of hidden spaces Rᵈ is structurally misaligned with the polar nature of semantic representations, a fact supported empirically by the anisotropy of neural activations and by the recent success of polar quantization schemes for KV caches.
In this paper we introduce φ-dynamics as a new, theoretically grounded inductive bias for LLM architectures. We establish three central results. First, OOD robustness requires minimizing the topological complexity of the hidden-state trajectory, as measured by Betti numbers. Second, the logarithmic spiral is the unique scale-invariant curve in R²ᵏ with a monotone phase. Third, the golden ratio φ = (1+√5)/2 is the uniquely optimal phase-shift parameter in the sense of Hurwitz’s theorem on Diophantine approximation. Building on these results, we introduce a differentiable regularizer Lφ that can be added to any existing architecture without structural modification, and we propose a concrete experimental verification protocol.
Files
phi_dynamics_EN-1.pdf
Files
(193.5 kB)
| Name | Size | Download all |
|---|---|---|
|
md5:1b9ecb4f753a21a53b38dc91db0405ba
|
193.5 kB | Preview Download |
Additional details
Dates
- Other
-
2026-04-01
References
- 1. Kim, L., Kim, S. Topological invariants of formal systems. Zenodo, 2026. DOI: 10.5281/zenodo.18907501. 2. Kim, L., Kim, S. Topological invariants of formal systems (Continuation). Zenodo, 2026. DOI: 10.5281/zenodo.18917389. 3. Kim, L., Kim, S. Equivalence of Algorithmic and Topological Complexity in Non-Boolean Spaces. Zenodo, 2026. DOI: 10.5281/zenodo.19080050. 4. Ethayarajh, K. How Contextual are Contextualized Word Representations? Comparing the Geometry of BERT, ELMo, and GPT-2 Embeddings. EMNLP, 2019. 5. Hooper, C. et al. KVQuant: Towards 10 Million Context Length LLM Inference with KV Cache Quantization. arXiv:2401.18079, 2024. 6. Ashkboos, S. et al. QuaRot: Outlier-Free 4-Bit Inference in Rotated LLMs. arXiv:2404.00456, 2025. 7. Su, J. et al. RoFormer: Enhanced Transformer with Rotary Position Embedding. arXiv:2104.09864, 2021. 8. Gu, A., Dao, T. Mamba: Linear-Time Sequence Modeling with Selective State Spaces. arXiv:2312.00752, 2023. 9. Hurwitz, A. Ueber die angenäherte Darstellung der Irrationalzahlen durch rationale Brüche. Mathematische Annalen, 39(2):279–284, 1891. 10. Hardy, G.H., Wright, E.M. An Introduction to the Theory of Numbers. 6th ed. Oxford University Press, 1979. 11. Kuipers, L., Niederreiter, H. Uniform Distribution of Sequences. Wiley, 1974. 12. Halton, J.H. On the efficiency of certain quasi-random sequences of points in evaluating multi-dimensional integrals. Numerische Mathematik, 2:84–90, 1960. 13. Edelsbrunner, H., Letscher, D., Zomorodian, A. Topological Persistence and Simplification. Discrete and Computational Geometry, 28:511–533, 2002. 14. Gebhart, T. et al. Characterizing the Shape of Activation Space in Deep Neural Networks. ICMLA, 2019. 15. Rieck, B. et al. Neural Persistence: A Complexity Measure for Deep Neural Networks Using Algebraic Topology. ICLR, 2019. 16. Kaplan, J. et al. Scaling Laws for Neural Language Models. arXiv:2001.08361, 2020. 17. Tishby, N., Schwartz-Ziv, R. Opening the Black Box of Deep Neural Networks via Information. arXiv:1703.00810, 2017. 18. Douady, S., Couder, Y. Phyllotaxis as a Physical Self-Organized Growth Process. Physical Review Letters, 68(13):2098–2101, 1992. 19. Berger, M. Geometry I & II. Springer, 1987.