Published February 18, 2026 | Version v1
Preprint Open

Sleeping LLM: Two-Phase Memory Consolidation for Lifelong Learning from 3B to 70B Parameters

Authors/Creators

  • 1. Independent

Description

We present a two-phase sleep architecture for memory consolidation in language models, with slow-wave sleep (SWS) for individual fact consolidation via per-fact LoRA training, and REM sleep for knowledge integration via synthetic multi-fact conversations. We introduce per-fact staged consolidation where each fact independently advances through stages (0-3) based on individual chat recall testing, replacing all-or-nothing per-edit gating. Key findings: MEMIT achieves near-zero perplexity cost for fact injection; REM reduces SWS-induced perplexity damage by 88% at 3B; per-fact gating achieves 95% consolidation success at 8B; and we discover pathway separation where MEMIT edits the raw completion pathway while LoRA edits the chat pathway. We validate across 3B, 8B, and 70B models, demonstrating that the graduated MEMIT dissolution schedule (scale 1.0 -> 0.5 -> 0.1 -> 0.0) successfully transfers knowledge from MEMIT to LoRA.

Notes

Part of the Sleeping LLM research series on sleep-wake memory consolidation for lifelong learning in language models.

Files

4-Sleeping-LLM.pdf

Files (123.6 kB)

Name Size Download all
md5:52336b0ed005e5618e25ca8f1a022ef3
123.6 kB Preview Download

Additional details

Related works

Continues
Preprint: 10.5281/zenodo.18778764 (DOI)
Is continued by
Preprint: 10.5281/zenodo.18778768 (DOI)