A Taxonomy of AI Governance Approaches: Distinguishing Visibility, Alignment, and Authorization
Description
The term “AI governance” has become semantically overloaded, applied indiscriminately to logging, guardrails, model alignment, and policy workflows. This paper proposes a formal taxonomy that distinguishes three fundamentally different governance problems—operational visibility, behavioral alignment, and decision authorization—and maps prevalent governance approaches to the problems they actually solve.
The paper introduces a precise definition of deterministic AI governance as a pre-execution authorization layer in which identical governed state produces identical governance verdicts and where the system emits verifiable decision artifacts sufficient for independent third-party replay. It provides a normative specification of governed state, a minimum evidence package, an enforcement triad (ALLOW / DENY / ABSTAIN), and practical evaluation frameworks—including anti-laundering tests designed to prevent trust-based or vendor-dependent imitation.
The taxonomy is intentionally testable and disqualifying rather than aspirational. It is intended to provide enterprise buyers, regulators, researchers, and auditors with a shared vocabulary and concrete criteria for evaluating AI governance claims in high-stakes, regulated environments.
Files
AI_Governance_Taxonomy_Paper_v1_4.pdf
Files
(465.5 kB)
| Name | Size | Download all |
|---|---|---|
|
md5:df14bc6652f8cd4ab17a0dcf8b101766
|
465.5 kB | Preview Download |
Additional details
Related works
- Is new version of
- Other: 10.5281/zenodo.18275970 (DOI)
Dates
- Issued
-
2026-01-17Initial public release of Version 1.4