Human-inspired AI


Towards AI with cognitive-dissonance and novelty-aware continual learning of episodic events

🚧 Work in Progress 🚧

I believe that Large Language Models (LLMs) led us to discover a powerful approach to implement conceptual simulation over possible worlds. I believe that the fundamental technology exists and argue that only a bit more work is needed to align these models with our world. This year, I began focusing on episodic memory as a key missing ability in current AI systems. This page outlines my approach and serves as a placeholder for ongoing work.

1. Why are LLMs so powerful?

My answer is: conceptual simulation over possible worlds

Navigating through possible worlds

From a given context, a Large World Model can sample possible futures, some are "real" and some are "imaginary"

Large language models have shown remarkable ability in building world models (multi-modal models are likely to be even more effective in this regard). Their power lies in their capacity to produce plausible outputs given a context, allowing them to explore possibilities and even creatively combine unrelated concepts.

2. Two main drawbacks:

The two big problems of current LLMs

The two major problems of current Large World Models

When LLMs “generate possible futures”, they struggle to distinguish between what is real and what is not. Additionally, while humans engage in step-by-step trial and error processes, LLMs can do this today but forget beyond their context window.

3. A “hippocampus” is a key missing piece towards AGI

3.1. Episodic memory is the solution to confabulations

Star wars in a box man

The star-wars-in-a-box man

Imagine a person locked inside a closed box, and who has only seen Star Wars movies since its birth. For this individual, wouldn’t Star Wars be their reality? Similarly, LLMs are trained on both fiction and reality without necessarily an explicit, native distinction between the two. They can recite knowledge, but they don’t know how and where they acquired it.

Episodic grounding into space and time are what make real events real.

I propose a simple, perhaps bold claim:

Grounding in space and time is what allows us to distinguish between what is real and what is wrong. Everything that happened in our space and time is real, and everything else is imaginary. The cue of space and time is what allows us to simply distinguish between both.

Unlike LLMs, human learning is constantly grounded in continuous space and time signals. We always know our location, date, and time while we think, learn and recall memories.

Hippocompass

The hippocampus should have been better called the hippocompass! Somewhat our coordinate system in time and space, and the corresponding index with which we can access our memories.

3.2. Episodic memory is also the solution to coherent system 2 thinking

I also think that episodic memory is crucial for enabling what is called “system 2” thinking in AI systems.

An episodic memory augmented LLM

A schematic view of how a system 2 thinking approach could be enabled if we solve the problem of ingesting episodic memory inside Large World Models. A similar approach could be implemented with today's technology/engineering, but would be too costly.

If we can quickly update LLM parameters coherently with the results of the intermediate thinking steps, we can enable true system 2 thinking, allowing models to remember and build upon their recent thoughts and actions.

4. Current work-in-progress

As the principal investigator, I crafted the research framework and guided the project from its inception to execution. However, the concrete progress we’re making wouldn’t be possible without the invaluable contributions of my dedicated colleagues, whose insights continue to enrich and refine these ideas. Below is what we will deliver soon.

🔜 Coming Soon: Episodic-memory benchmarks

A benchmark to evaluate LLMs on their ability to recall episodic events. This involves a generic framework to represent episodic events that happen in time and space, and a systematic list of episodic recall tasks.

Stay tuned!

🔜 Coming Soon: Empirical investigations on cognitive dissonance-aware incremental update of knowledge inside LLMs.

An LLM blindly believes its training data even if it contradicts its prior knowledge, while a human detects when something contradicts their prior beliefs. When such cognitive dissonance occurs, the human actively engages in resolving it. Additionally, implementing a form of stubbornness, where neurons full of knowledge become stubborn, and weights that don't activate clearly enough are reset to zero, could be beneficial.

Stay tuned!