Simplifying Peer Review for Accessibility: A Case Study on GPT 4's Performance in Finance Using Cognitive-Informed Prompts
Description
This study investigates how GPT-4 performs in simplifying peer reviewer comments in the field of academic finance, particularly for cognitively diverse users such as those with dyslexia or working memory challenges. Using ten domain-specific reviewer comments and two types of prompts—one general and one tailored for cognitive accessibility—the paper evaluates 40 GPT-4 outputs for accuracy, consistency, and conceptual fidelity. The results reveal frequent misrepresentations of key technical terms, oversimplifications that distort meaning, and inconsistency across runs, regardless of prompt type. The findings highlight critical limitations in using large language models for accessibility-focused simplification in specialized academic domains and call for more domain-sensitive AI development.
Files
Simplifying Peer Review for Accessibility_A Case Study on GPT 4’s Performance in Finance Using Cognitive-Informed Prompts.pdf
Files
(327.9 kB)
| Name | Size | Download all |
|---|---|---|
|
md5:3c2ac8a28898278922e827847c39ab38
|
327.9 kB | Preview Download |
Additional details
References
- OpenAI. (n.d.). GPT 4. In Wikipedia. Retrieved June 2025, from https://en.wikipedia.org/wiki
- Wang, Z., et al. (2023). Can LLMs like GPT 4 outperform traditional AI tools in dementia diagnosis? arXiv. https://arxiv.org/pdf/2306.01499
- Wu, X., & Arase, Y. (2024). An in depth evaluation of GPT 4 in sentence simplification with error based human assessment. arXiv. https://arxiv.org/abs/2403.04963
- Picton, B., et al. (2024, November). GPT 4 Improves Readability of Institutional Heart Failure Patient Education Materials: An Observational Study. medRxiv. https://www.sciencedirect.com/science/article/pii/S1386505624004064