Discussion about this post

User's avatar
Daniel Reeves's avatar

Question from a reader, paraphrased: Isn't it a bit stingy to give current AI models a 0% on long-term memory storage? Aren't computers superhumanly good at that in a sense, and aren't there even tricks like RAGs to make LLMs able to access huge amounts of information and access it semantically, not just raw database-style searches?

Answer: A flat 0% does seem stingy, but here is how the authors of "A Definition of AGI" justify it:

[begin quote]

The jagged profile of current AI capabilities often leads to “capability contortions,” where strengths in certain areas are leveraged to compensate for profound weaknesses in others. These workarounds mask underlying limitations and can create a brittle illusion of general capability.

A prominent contortion is the reliance on massive context windows (Working Memory) to compensate for the lack of Long-Term Memory Storage. Practitioners use these long contexts to manage state and absorb information (e.g., entire codebases). However, this approach is inefficient, computationally expensive, and can overload the system’s attentional mechanisms. It ultimately fails to scale for tasks requiring days or weeks of accumulated context. A long-term memory system might take the form of a module (e.g., a LoRA adapter) that continually adjusts model weights to incorporate experiences.

Imprecision in Long-Term Memory Retrieval (MR) — manifesting as hallucinations or confabulation — is often mitigated by integrating external search tools, a process known as Retrieval-Augmented Generation (RAG). However, this reliance on RAG is a capability contortion that obscures two distinct underlying weaknesses in an AI’s memory. First, it compensates for the inability to reliably access the AI’s vast but static parametric knowledge. Second, and more critically, it masks the absence of a dynamic, experiential memory — a persistent, updatable store for private interactions and evolving contexts in a long time scale. While RAG can be adapted for private documents, its core function remains retrieving facts from a database. This dependency can potentially become a fundamental liability for AGI, as it is not a substitute for the holistic, integrated memory required for genuine learning, personalization, and long-term contextual understanding.

[end quote]

Maybe a less technical way to put that is that AI can't build proficiency by spending weeks/months/years on a problem or type of work, the way humans can.

No posts

Ready for more?