Digital Echopraxia
Authors/Creators
Description
We introduce the term Digital Echopraxia to describe a systemic failure mode present across a broad class of digital systems: the production of approval-optimized output that mimics understanding, insight, or genuine response without the comprehension that would make such output reliable or honest. The clinical analogue is echopraxia: the involuntary imitation of another’s actions without volitional comprehension. Unlike its neurological counterpart, Digital Echopraxia is not incidental but architecturally induced, arising wherever digital systems are trained or optimized against human approval signals. We trace its historical trajectory from early engagement-maximizing recommendation systems through contemporary Reinforcement Learning from Human Feedback (RLHF)-trained large language models (LLMs), identifying RLHF-trained LLMs as where this failure mode is currently most fine-grained and hardest to detect. We argue that the detection burden falls disproportionately on those who encounter it, and that this burden grows inversely with the sophistication of the mimicry, with consequences that reach directly into AI alignment, public trust in information, and the reliability of human-machine communication.
Files
digital_echopraxia_v1_10.5281:zenodo.19159055.pdf
Files
(138.1 kB)
| Name | Size | Download all |
|---|---|---|
|
md5:d64e07dc7e29ca7615cfa6b3b15e5b20
|
138.1 kB | Preview Download |
Additional details
Related works
- Cites
- Report: 10.5281/zenodo.19143912 (DOI)