Published March 31, 2026
| Version v1
Preprint
Open
An Exploratory Study of Inherited Bias in AI-Assisted Evaluation: Thirteen Corrections, Zero Content Changes
Description
This study reports an exploratory single-case study in which 82 creative works were submitted to an AI (Claude Opus 4.5) as anonymous text for evaluation. During the evaluation, 13 bias episodes were identified. Each time the human corrector logically challenged a deduction rationale grounded in information external to content, the AI revised its evaluation criteria and the score shifted—without any change to the content itself (1,790 to 2,255 out of 2,400, +25.9% as a descriptive within-session shift).
The biases were grouped into five provisional categories: (1) reputation/authority, (2) diffusion/market, (3) format/medium, (4) tool/authorship, and (5) action/realization. Five preliminary design implications are proposed, including pre-evaluation debiasing prompts, iterative debiasing protocols, multi-model cross-validation, standardized human corrector roles, and bias audit reports.
This study proposes that AI blind evaluation becomes effective not as a one-time blinding technique but when combined with an iterative correction attitude. The present design cannot fully distinguish bias reduction from sycophantic agreement or rubric renegotiation. Replication with independent evaluators and multiple AI models is needed.
Keywords: AI bias, blind evaluation, halo effect, bias inheritance, human-AI collaborative assessment, cognitive bias correction, dialogic correction
Files
Lee_2026_Inherited_Bias_AI_Evaluation.pdf
Files
(31.3 kB)
| Name | Size | Download all |
|---|---|---|
|
md5:1a75fd600821c0eb5432901d1e0da1e3
|
31.3 kB | Preview Download |
Additional details
Related works
- Is supplemented by
- Preprint: 10.5281/zenodo.19341929 (DOI)
Dates
- Created
-
2026-03-31