The Generative Brain

The Generative Brain

The brain is a non-Markovian dynamical system. Each cognitive state is the optimized generative outcome of the states that preceded it.

The Autoregressive Brain framework proposes that cognition operates as sequential state generation — not by storing and retrieving representations, but by continuously generating each next state as the optimal continuation of the preceding trajectory. Memory, language, reasoning, and consciousness are all downstream consequences of this single principle.

Large language models serve as existence proofs: autoregressive prediction over learned distributional structure produces composition, context-sensitivity, and apparent understanding — without any of the machinery cognitive science has traditionally assumed is required.

The Autoregressive Brain

A new framework for cognition, memory, and language

Generation, Not Retrieval

Remembering is reconstructive generation, not playback from storage. Memory is the capacity to regenerate experiential trajectories, constrained by learned structure.

Syntax as Distributional Epiphenomenon

What linguists call "grammar" is not a rule system but the skeletal structure of distributional statistics. Function words are high-leverage tokens that constrain sequential prediction.

Consciousness as Recursive Autoregression

Subjective experience emerges from the cognitive system modeling its own generative process. The "self" is a trajectory through state-space, not an entity.

LLMs as Existence Proofs

Large language models demonstrate that composition, apparent reasoning, and memory-like behavior can emerge from autoregressive prediction without explicit storage or rule systems.

Read the Full Theory →

Latest Writing

From the blog

  • Syntax Without Rules: The Distributional Skeleton of Language

    The most consequential finding in modern linguistics did not come from a linguistics department. It came from scaling next-token prediction: large language models acquire grammatical competence — subject-verb agreement across long dependencies, recursive embedding, tense consistency, productive morphology, pragmatic sensitivity...

  • Memory as Trajectory Regeneration

    The storage-retrieval model of memory has been the default framework in cognitive science since Ebbinghaus. Experiences are encoded, consolidated into stable representations, and later retrieved. The metaphor shifts with each generation — wax tablets, filing cabinets, hard drives — but...

  • World Properties Without World Models: What Word Embeddings Know About Geography

    A central claim of the autoregressive brain framework is that distributional structure learned over sequential experience is sufficient to support contextually appropriate generation about the world — without the system needing to maintain explicit world models, stored representations of physical...

All Posts →

Videos

Conversations and presentations on the autoregressive mind

The (Disturbing) Theory of Autoregressive Minds

Theories of Everything with Curt Jaimungal

Discussion with Michael Levin

Thoughtforms Life

All Videos →