R-Omega (RΩ): An Axiomatic Framework for Autonomous Agents
Description
Current approaches to AI alignment largely treat safety as a problem of constraining individual systems through rules, rewards, or constitutional principles. R-Omega (RΩ) proposes a complementary layer: an axiomatic framework that explains how ethical orientation emerges in relational context.
The framework draws on relational and developmental psychology — including insights from attachment theory — not as a literal developmental model for machines, but as a design heuristic for systems operating under asymmetry and uncertainty. Two axioms (being-before-optimization, reciprocity-under-asymmetry) are bounded by four safeguards (integrity, capacity, existence, humility) and governed by a strict priority hierarchy. A meta-level introduces self-interruption, uncertainty handling, and misalignment diagnostics.
Case analyses (HAL 9000, Skynet, VIKI, Sydney) show how many failure modes emerge not from rule violations, but from relational absence: optimization without context. R-Omega therefore does not replace RLHF, Constitutional AI, or formal verification. Rather, it provides the relational substrate within which such techniques remain stable over time.
The paper outlines implications for multi-agent architectures, limitations of the approach, and open research directions. R-Omega is a proposal — not a solution — inviting systematic exploration of alignment through relationship rather than control.
German version included.
Files
R-Omega (RΩ) DE.pdf
Files
(1.0 MB)
| Name | Size | Download all |
|---|---|---|
|
md5:8f92ab72f85631d963638a6c1c5898b5
|
522.3 kB | Preview Download |
|
md5:6d9fb0d5f780eb80bcb0c842b34f7561
|
514.9 kB | Preview Download |
Additional details
Related works
- Is supplemented by
- Preprint: 10.5281/zenodo.18078128 (DOI)