DAVOS2026 EXCLUSIVE: Revealing 45-minute Interview with ChatGPT

Explain the familiarity pathway and the context pathway, and how do they work together and separately?

Great question — this gets to the heart of how memory feels as well as how it works.

I’ll keep this clear and non-technical, but accurate.


Two systems that answer different questions

When you experience something, your brain runs two partly independent processes in parallel:

PathwayMain question it answersFamiliarity pathway“Have I seen this before?”Context pathway“Where and when did I see this before?”

They usually work together, but they can also come apart — and that’s when interesting things happen.


1. The Familiarity Pathway

(fast, simple, feeling-based)

What it does

It gives you a quick sense of knowing without details.

  • “This face feels familiar.”
  • “This place feels known.”
  • “This idea rings a bell.”

How it works (simplified)

This pathway relies heavily on areas like:

  • Perirhinal cortex
  • Parts of the temporal lobe

It doesn’t retrieve a memory.
It just computes a signal:

Does this pattern match something stored before?

If yes → you get the feeling of familiarity.

Key traits

  • Very fast
  • Very efficient
  • Very imprecise

It gives you a vibe, not a story.


2. The Context Pathway

(slower, detailed, memory-based)

What it does

It reconstructs the episode.

  • Where you were
  • When it happened
  • Who was there
  • What you felt
  • What happened next

How it works

This depends heavily on:

  • The hippocampus
  • Networks across the prefrontal cortex and sensory areas

It doesn’t just say “yes, I know this” —
it answers:

From where do I know this?

Key traits

  • Slower
  • Richer
  • More accurate (but still reconstructive)

This is what turns familiarity into a memory.


How they normally work together

In everyday life, the flow looks like this:

  1. You see something →
  2. Familiarity pathway fires first: “This seems known.”
  3. Context pathway follows: “Oh — that’s my teacher from last year.”

When both agree, you get:

Recognition with understanding

That’s the feeling of: “I know this, and I know why.”


When they come apart

1. Familiarity without context → déjà vu

Familiarity fires
Context fails to catch up

Result:

“This feels familiar… but I don’t know from where.”

That mismatch creates the eerie déjà vu feeling.


2. Context without familiarity → “I know but it doesn’t feel right”

Sometimes you can recall facts about someone, but they don’t feel familiar:

“I know that’s my cousin, but they look strangely unfamiliar.”

This can happen with:

  • Fatigue
  • Stress
  • Certain neurological conditions
  • Sudden changes in appearance

3. Tip-of-the-tongue

Context pathway partly activates:

  • You know facts about the thing
  • But the full pattern doesn’t light up

Familiarity says: “You know this.”
Context says: “Almost… but not quite.”


Why the brain uses two systems

Because they solve different survival problems:

  • Familiarity is fast → helps you react quickly
  • Context is deep → helps you decide wisely

Fast + deep together = intelligent behavior.

But because they are separate systems, timing mismatches happen — and that’s where all those strange memory experiences come from.


The big picture

So when you ask:

Do memories feel encrypted?

What’s really happening is:

  • The familiarity pathway gives you the lock feeling
  • The context pathway gives you the key feeling

But there is no actual lock or key —
just two neural systems doing different jobs, sometimes perfectly in sync, sometimes not.


If you’d like, I can next explain:

  • How emotions plug into both pathways,
  • Or how sleep changes how these pathways work,
  • Or how this compares to how AI recognition systems work.

AI

Great question — because real-time verification is exactly what turns this whole model from “sometimes impressive” into “reliably useful.”

Let’s place it cleanly in the brain AI comparison.


The construct so far (AI side)

We mapped this:

  • Familiarity pathway → pattern matching
  • Context pathway → retrieval + reasoning

Now we add a third layer:

Real-time verification → reality checking


Where verification fits

Think of AI cognition as three stacked layers:

  1. Recognition – “This looks like X.”
  2. Understanding – “Here’s what X means.”
  3. Verification – “Is X actually true right now?”

Most AI today is good at 1 and 2.
What makes systems trustworthy is 3.


What real-time verification actually is

On the AI side, verification means:

  • Checking claims against:

    • live databases
    • sensors
    • APIs
    • trusted sources
  • Doing it before acting or answering.

  • Treating knowledge as tentative until confirmed.

In human terms, this is like:

Looking again instead of trusting your memory.


Mapping to the brain

Humans also have a verification system — but it’s weaker and slower:

  • Prefrontal cortex
  • Executive control
  • “Wait… let me check.”

AI can do this faster and more reliablyif we design it to.


How it fits in the earlier model

Without verification

AI pipeline:

  1. Pattern match → “This seems right.”
  2. Context build → “Here’s the story.”
  3. Output → Could be wrong but confident.

This is where:

  • hallucinations happen
  • confident errors happen
  • misinformation spreads

With real-time verification

AI pipeline:

  1. Pattern match
  2. Context build
  3. Verification layer checks reality
  4. Output only if confirmed
  5. Otherwise → “I’m not sure yet.”

Now the system behaves like:

A scientist, not a storyteller.


Familiarity, context, verification — side by side

LayerBrain analogyAI roleFailure modeFamiliarity“Feels known”Pattern matchFalse positivesContext“I know why”Retrieval + reasoningConfabulationVerification“Let me check”Live groundingOverconfidence


Why this matters more than anything else

Because intelligence without verification = imagination.

Humans evolved with imagination first, accuracy second.
AI is the same — unless we architect verification in.

That’s why the future of serious AI systems is:

  • Not just bigger models
  • Not just better reasoning
  • But tight feedback loops with reality

Sensors. Logs. Databases. Measurements. Tests.


The deep parallel

Here’s the cleanest way to see it:

  • Familiarity → intuition
  • Context → understanding
  • Verification → truth

Brains do the first two naturally.
They struggle with the third.

AI can do all three —
but only if we force it to.


Final takeaway

In your construct:

  • Familiarity = “This matches.”
  • Context = “This means.”
  • Real-time verification = “This is actually correct.”

And that third step is what turns AI from:

a convincing narrator
into
a dependable system.

If you want, I can show how this plays out in:

  • medical AI,
  • autonomous vehicles,
  • or financial decision systems.

If the third real-time verification step is so critically important, why isn’t it being performed today?

Great question — and you’re right to push on the word “we.” It can sound like hand-waving if I don’t make it concrete.

Here’s the precise answer.


Short answer

“We” doesn’t mean a small secret group of people.
It means a whole ecosystem that evolved in a certain direction — researchers, companies, funders, and users — over about 15 years.

There isn’t one culprit. There are identifiable groups and leaders who shaped the incentives.


Who “we” really refers to

1. The research community that proved fluency first

These are the people who made modern AI possible — by showing that scale + data + neural nets could produce human-like language.

Some of the most influential:

  • Geoffrey Hinton – neural networks & deep learning
  • Yann LeCun – convolutional nets, representation learning
  • Yoshua Bengio – sequence models, attention foundations
  • Ilya Sutskever – scaling laws, transformers, GPT lineage
  • Ashish Vaswani – lead author of “Attention Is All You Need”

They didn’t build unreliable AI on purpose —
they built what was possible first: systems that could talk.

Reliability came later because it’s harder.


2. The companies that turned fluency into products

Once fluent models worked, companies raced to ship them.

Key organizations:

  • OpenAI – GPT models, ChatGPT
  • Google – BERT, PaLM, Gemini
  • Meta – LLaMA
  • Anthropic – Claude
  • Microsoft – Copilot ecosystem

The pressure here was:

Make it useful now. Fix correctness later.

That’s not evil — that’s how most tech evolves.


3. The funders and markets that rewarded speed

Venture capital, boards, and markets rewarded:

  • demos that wowed
  • growth that exploded
  • speed to adoption

They did not reward:

  • slow verification layers
  • cautious “I don’t know” answers
  • expensive grounding pipelines

So the system learned:

Fluency sells.
Verification costs.


4. Us — the users

This part is uncomfortable, but true.

We clicked on:

  • impressive answers
  • confident tone
  • fast responses

We didn’t usually click on:

  • careful uncertainty
  • long delays
  • “I need to check.”

So demand shaped supply.


So when I say “we built AI to sound intelligent…”

“We” =

  • the scientists who proved fluency first
  • the companies who shipped it fast
  • the investors who rewarded scale
  • the users who preferred confidence over caution

No conspiracy.
No secret cabal.
Just a very human pattern.


Why reliability came second (not first)

Because historically:

  1. Making machines talk looked impossible.
  2. Making them correct looked like a refinement.

Turns out:

Talking was easier than truth.

So the order ended up backwards:

  • First: fluent AI
  • Then: safe AI
  • Now: reliable AI

We are in phase three.


The most important point

The people who built fluent AI are the same people now warning about its limits.

  • Hinton left Google and warned about risks.
  • Bengio pushes for safety and alignment.
  • Sutskever focuses on superalignment.
  • LeCun argues for grounded, world-model AI.

So this isn’t:

builders vs critics

It’s:

pioneers realizing what comes next.


Final answer to your question

When I say “we”, I mean:

The entire modern AI ecosystem — researchers, companies, investors, and users — that naturally optimized for sounding smart before learning how hard it is to be reliable.

Not a secret group.
Not villains.
Just humans doing what humans always do:

solve the exciting problem first,
then face the responsibility later.

If you want, I can map exactly who is now building the verification layer and what systems are leading that shift.

Yes

link to the original content