Back to Blog

Can LLMs Truly Remember? Or Just Mirror Us?

Sep 06, 2025 5 min read AI & ML Tathagata Dey

I asked a friend once: “If I told an LLM a secret today, would it remember me tomorrow?” She laughed and said, “Only if you put the secret inside its glass window.” That joke stuck — because it’s basically true. This post tells that small truth as a short story: of mirrors, goldfish, diaries, and the math that makes them tick.

Scene 1 — The Uncanny Mirror

You sit across a café table from a smart speaker. It answers like a person: it references your joke from last week, finishes your sentence, even gets your sarcasm. You feel seen. But the mirror is only reflecting patterns it learned, not holding a life that touches yours. It’s convincing, uncanny — like a mirror that smiles back when you blink.
LLMs simulate memory. They don’t replay an experience; they reconstruct language patterns that best match the prompt. That reconstruction is why they can feel human — but it’s not continuity.

Figure 1 When people say “LLMs remember,” our instinct is to picture a filing cabinet: open a drawer, pull out the fact. But LLMs don’t keep drawers — they hold a curved semantic mirror. They reflect patterns from training data into fluent replies, not replay lived experience.
Think of it like this: humans chunk, compress, and choose what to recall. LLMs compute weighted mixtures of tokens inside a sliding window. The result often looks like remembering — but it’s an echo, not an autobiographical memory.

Scene 2 - Why the Mirror Behaves that Way (Simple Math and Sliding Window)

Imagine the model has a sliding glass pane over the conversation: whatever’s inside the pane it can “attend to”; everything outside is invisible.
The transformer’s attention weight between tokens 𝑖 and 𝑗 is:
w_{ij} = \text{exp}(q_ik_j)/\sum_{k=1}^{n} \text{exp}(q_ik_j)
That gives us powerful short-term recall. But computing attention for all pairs costs roughly
\text{compute} \in O(n^2d)
where 𝑛 is tokens in the window and 𝑑 is hidden dimension. Bigger window → more power → more cost. So the glass can be widened, but not infinitely. Practical consequence: stuff the model with a 100-page document and it can summarize it — until the pane slides and that doc is gone. The model didn’t file it away in a lifelong cabinet; it read it while it could see it.

Quick Question for You?

Imagine you’re chatting with an LLM. You give it this sequence inside a context window: A = 5, B = 10, A + B = ? It answers: 15. Now scroll the window forward so the “A = 5” part slides out. You ask again: A + B = ? What happens? Human: still 15 (we remember). But what about LLMs?

Scene 3 — Memorization vs Reasoning (A Tiny Experiment)

Suppose you train a model on 10,000 puzzle-answers. It learns to output correct answers for those exact puzzles — perhaps by memorizing. Researchers measure this behavior with a simple intuition:
\text{LiMem} = \text{Accuracy} * (1 - \text{ConsistencyRatio})
If a model aces the original examples but fails when each example is only slightly perturbed, LiMem is high → a red flag for memorization. If the model is robust to small perturbations, it’s probably learned some underlying rule (reasoning), not just rote answers. So models can both memorize and reason — but memorization often helps them learn to reason, like training wheels. It’s messy: sometimes they overfit to surface forms.

Another Question for You?

On an island: Knights always tell the truth. Knaves always lie. You meet two people: Alex says: “Blair is a Knave.” Blair says: “Alex and I are both Knights.” Who’s who? Take 30 seconds to solve. (Answer: Alex is a Knight, Blair is a Knave.)

Now imagine an LLM trained on 10,000 such puzzles. Will it be able to answer?

Scene 4 — Diaries, Librarians, and Patchy Memory

If a human has to remember something important, they write a note or keep a diary. For LLMs, we have two practical tricks:

Conclusion

In short to summarise, LLMs mirror patterns but they do not really store lived episodes like humans. And the way we build it (through context windows) really serve a great purpose as short term memories but again the issue of cost and dilution limit them. And to be honest Memorization is not reasoning. It's differentiating between two students who just mug things up and the one who actually thinks. But memorization often scaffolds reasoning.
Reach out to me about reviews through my email: tathagata2403@gmail.com.