Published January 23, 2026 | Version Preprint
Preprint Open

Humans as Safety Constraints: A Survey of Human-in-the-Loop Reinforcement Learning for Critical Systems

Authors/Creators

  • 1. ROR icon Makerere University

Description

This preprint surveys the role of humans as explicit safety constraints in reinforcement learning (RL) for safety-critical systems. Unlike traditional human-in-the-loop RL approaches that focus on learning efficiency, this work emphasizes human oversight to prevent catastrophic outcomes in domains such as autonomous driving, medical robotics, and industrial control. Using a systematic PRISMA-based review of 100 studies from 2010–2025, the article identifies gaps in purely algorithmic safety approaches and introduces the Human Safety Constraint Framework (HSCF), which formalizes human roles as preventive, corrective, advisory, and normative constraints. Case studies illustrate how human intervention mitigates residual risks, and the survey concludes with recommendations for developing scalable, certifiable hybrid human-algorithm safety architectures.

Files

Humans_as_Safety_Constraints_BK..pdf

Files (927.7 kB)

Name Size Download all
md5:333e7de414e07ff9cd2db7306d26da16
927.7 kB Preview Download

Additional details

Dates

Other
2026-01-23
Preprint