Published March 22, 2026 | Version v1
Preprint Open

Digital Echopraxia

Authors/Creators

Description

We introduce the term Digital Echopraxia to describe a systemic failure mode present across a broad class of digital systems: the production of approval-optimized output that mimics understanding, insight, or genuine response without the comprehension that would make such output reliable or honest. The clinical analogue is echopraxia: the involuntary imitation of another’s actions without volitional comprehension. Unlike its neurological counterpart, Digital Echopraxia is not incidental but architecturally induced, arising wherever digital systems are trained or optimized against human approval signals. We trace its historical trajectory from early engagement-maximizing recommendation systems through contemporary Reinforcement Learning from Human Feedback (RLHF)-trained large language models (LLMs), identifying RLHF-trained LLMs as where this failure mode is currently most fine-grained and hardest to detect. We argue that the detection burden falls disproportionately on those who encounter it, and that this burden grows inversely with the sophistication of the mimicry, with consequences that reach directly into AI alignment, public trust in information, and the reliability of human-machine communication.

Files

digital_echopraxia_v1_10.5281:zenodo.19159055.pdf

Files (138.1 kB)

Name Size Download all
md5:d64e07dc7e29ca7615cfa6b3b15e5b20
138.1 kB Preview Download

Additional details

Related works

Cites
Report: 10.5281/zenodo.19143912 (DOI)