There is a newer version of the record available.

Published December 24, 2025 | Version v2
Preprint Open

RAG: Source-Grounding Does Not Prevent Hallucinations in NotebookLM

  • 1. PatternPulseAI

Description

We posit that hallucinations are not failures of knowledge access but failures of semantic authority; retrieval systems constrain information sources but do not arbitrate meaning, and therefore cannot prevent hallucinations arising from unresolved semantic competition.This paper provides evidence that even with source-grounding, RAG will result in hallucinations despite marketing claims, regardless of source data, due to ambiguity and lack of semantic authority. 

 

Google markets NotebookLM as reducing hallucinations through source-grounding, constraining model outputs to uploaded documents rather than general training knowledge. This architectural claim would represent a solution to a fundamental problem in language model deployment. We tested this claim by replicating controlled disambiguation experiments from prior work on semantic governance failures. Using identical test protocols applied to GPT, Claude, and Grok, we subjected NotebookLM to conditions that reliably induce hallucinations in frontier models: strict semantic dominance, where a single interpretation must be maintained globally despite conflicting local context.

 

NotebookLM exhibited identical hallucination patterns under strict dominance (inventing muddy financial institutions, canoe docking facilities, and vegetation growth on bank buildings) and identical recovery under revocable dominance (clean context-dependent disambiguation). Additionally, NotebookLM explicitly acknowledged using “general knowledge” rather than source-constrained definitions during disambiguation tasks, demonstrating it is not a closed corpus system.

 

Source-grounding constrains retrieval but does not introduce semantic governance primitives. Hallucinations caused by inability to prioritize and revoke competing interpretations persist regardless of source quality. The findings challenge foundational assumptions underlying Retrieval-Augmented Generation (RAG) architectures and enterprise AI deployment strategies predicated on source-grounding as a hallucination mitigation technique.

Files

Source Grounding and RAGFINAL .pdf

Files (538.6 kB)

Name Size Download all
md5:9a0a3be4ed59544f7657c8f7cd662807
538.6 kB Preview Download

Additional details

Related works

Is supplement to
Publication: 10.5281/zenodo.17929851 (DOI)
Publication: 10.5281/zenodo.17831839 (DOI)
Publication: 10.5281/zenodo.17871463 (DOI)
Publication: 10.5281/zenodo.17847869 (DOI)

Dates

Available
2025-12-24
Title change; this paper shows that NotebookLM is closed corpus and reduces hallucinations