Personal Tensor Memory
Authors/Creators
- 1. Department of Computer Science Engineering, Independent Researcher, Chennai (Tamil Nadu), India.
Description
Abstract: Large language models (LLMs) excel at general knowledge but struggle when they must remember the preferences, profile facts, and long-term context of a specific user—especially on constrained devices. We introduce Personal Tensor Memory (PTM), a privacy-preserving add-on that assigns every user a fixed-shape matrix, which the frozen backbone can query through one additional attention head. A nightly routine— Hebbian add + decay, norm clipping, slot merge/evict, and occasional orthogonal rotation—re‑organises information inside that matrix without changing its shape or touching billions of backbone weights. On synthetic concept‑drift streams and anonymised personal‑assistant logs, PTM matches kNN‑LM perplexity while needing only 5 % of its context window, and surpasses rank‑8 LoRA under few‑shot data—all using < 8 MB per user and < 1 s daily CPU on a smartphone.
Files
E110005050825.pdf
Files
(498.8 kB)
| Name | Size | Download all |
|---|---|---|
|
md5:e55821cb4512d17c7cd34b1d0d1d301d
|
498.8 kB | Preview Download |
Additional details
Identifiers
- DOI
- 10.54105/ijainn.E1100.05050825
- EISSN
- 2582-7626
Dates
- Accepted
-
2025-08-15Manuscript received on 25 June 2025 | First Revised Manuscript received on 21 July 2025 | Second Revised Manuscript received on 01 August 2025 | Manuscript Accepted on 15 August 2025 | Manuscript published on 30 August 2025.
References
- Hu, E. et al. "LoRA: Low‑Rank Adaptation of Large Language Models." ICLR 2022. DOI: http://doi.org/10.48550/arXiv.2106.09685
- Khandelwal, U. et al. "Generalization through Memorization: Nearest Neighbor Language Models." ICLR 2020. DOI: http://doi.org/10.48550/arXiv.1911.00172
- Guo, D. et al. "GraphCodeBERT: Pre‑Training Code Representations with Data Flow." ICLR 2021. DOI: http://doi.org/10.48550/arXiv.2009.08366
- Zaken, E. B. et al. "BitFit: Simple Parameter‑Efficient Fine‑Tuning for Transformer‑Based Masked Language‑Models." arXiv 2021. DOI: http://doi.org/10.48550/arXiv.2106.10199
- Munkhdalai, T., Trischler, A. "Meta Networks." ICML 2017. DOI: http://doi.org/10.48550/arXiv.1703.00837
- Ghorpade, M. et al. "Efficient Low‑Rank Adaptation via Randomized SVD." arXiv 2023 . DOI: http://doi.org/10.48550/arXiv.2306.06029
- Chu, X. & Zaniolo, C. "Selective and Efficient Reservoir Sampling for Data Streams." IEEE TKDE 2020. DOI: http://doi.org/10.1109/TKDE.2020.2988027
- Ramasesh, V. et al. "An Embedding View of Continual Learning." NeurIPS 2021. DOI: http://doi.org/10.48550/arXiv.2102.06253