================
== thinks.lol ==
================
thoughts, laughing

Memory makes computation universal, remember?

What CRISPR, o3, and the Memphis Plume Show Us About Intelligence.

TL;DR: Universal intelligence—the ability to solve any problem—depends on the ability to be true to yourself while remembering what you’ve thought, and nothing else.

proof: arXiv:2412.17794


The Memphis Plume

On a crisp morning in early December 2024, I was running by the Mississippi when I noticed a plume on the horizon:

A large industrial plume over south Memphis.

At first, I thought it was xAI’s “Colossus” supercomputer. However, the plume is mostly from the Tennessee Valley Authority’s Allen Combined Cycle Plant—a 1.1-gigawatt natural gas power plant which sits on the same site. Nevertheless, let’s take it as a vision of the future that lies on our current development path.

With 100,000 Nvidia H100 GPUs, xAI is believed to be the world’s largest AI supercomputer. It’s already drawing around 150 MW and substantial cooling water. For now, that’s only a fraction of Allen’s power footprint, but xAI’s plans call for scaling to a gigawatt in the coming year—potentially creating its own towering plume of the same scale in the race toward artificial general intelligence.

When that scale-up is complete, the Memphis skyline will feature a massive plume representing our rush to train an artificial general intelligence (AGI). xAI’s stated goal is a “1% shot at a Kardashev type 1 civilization”—a society capable of harnessing all its planet’s energy, presumably enabled by the AGI they aim to bring into the world. And they’re just getting started, with a tenfold expansion of Colossus to 1M GPUs already planned.

Universal Computation

You may have noticed that, although ravenous and hyperactive, a human child doesn’t generate a thousand meter plume of water vapor and reactive particulates to develop their intelligence. Yet your average human, fueled by sandwiches, can implement any algorithm with just pen and paper. So augmented with bronze age technology, we become “universal” computers in Turing’s language—the quintessential “general” intelligence capable of solving any problem given enough time and snacks. For the sake of argument, let’s agree: general intelligence ≈ universal computer.

That link has been on my mind for two years, ever since I first read the magic words “Let’s think step by step” in a groundbreaking machine learning paper. I was amused by how a simple phrase, first introduced in “Large Language Models are Zero-Shot Reasoners”, seemed to invoke a more deliberate, thoughtful mode in large language models (LLMs). It was as if some psychological effect flipped them into a more rigorous and structured part of their state space, leading to dramatic improvement in performance, and encouraging the world’s LLM whisperers to add this incantation to every time they saw models faltering.

But theoretical analyses of transformers revealed something deeper than just a neat psychology trick. Transformers, in a single forward pass, are restricted to a computational class called TC0, basically a family of threshold circuits that can be computed in parallel with aggregation. TC0 systems are often called “bounded parallel”. They can quickly match patterns, but can’t count, do math, or follow a recipe unless the process can be encoded in a fixed set of circuits. However, if you use “chain-of-thought” or a series of recursive prompts—feeding each of the model’s outputs back into its input—the combined system becomes capable of simulating a universal Turing machine. The model’s expressivity, harnessed across multiple steps, blossoms into general computation.

Memory Makes Computation Universal

Out of deep love for this line of research, I spent much of December in a meditation on the link between memory and computation. There and in this post, I provide the simplest and most intuitive explanation of universality (a.k.a. general intelligence) that I could synthesize:

Universality requires only (1) stable evolution of thought (no hallucinations), and (2) reliable access to the history of thought.

This synthesis rightly seems banal, dull, too simple to be useful, too theory of computation 101 to be interesting. And OK, if you studied the theory of computation, yes, it is, but why did y’all forget the textbook and start imagining that threshold circuits and feedforward networks could achieve general intelligence? If you can remember, this perspective quickly takes us to some amazing places.

The first place is an intuitive understanding of what LLMs are. Machine learning got excited by the observation that adding more parameters and training tokens could yield “scaling laws” that took us to a place where intelligence “emerged”. Another interpretation is that LLMs are a quintessential example of transfer learning, where humans and our society is the model that’s being transferred. By trying to predict what humans might say or do next, the models effectively learned patterns embedded in our writing—including the latent computational rules that shape human reasoning. They’re not just modeling words but the programs behind those words (which might be human minds, or other models, or entire cryptographic protocols). Looking through the lens of Turing machines, we’ve transferred our state transition functions to them, which they can now generalize to new input contexts (drawing tikz unicorns and such). This deeper transfer of computational thinking explains why models continue improving far beyond naive pattern-matching saturation points (technical paper)—they’re gradually refining their models of human cognitive processes themselves.

Part of the magic is that LLMs, most of which are Transformers, are forging a shared computational interface with us, letting us push each other around in a collaborative workspace that can solve new tasks.

The Bitter Pill

However, they have a costly flaw that makes them otherworldly in their resource demands: attention. In effect, attention is a learned function for memory and association across the input. But with n tokens, each must attend to every other token—an n × n explosion in both compute and memory. Double your context length, quadruple the cost. Want to “think” four times longer? That’s sixteen times more overhead. And because memory is the key to intelligence, model context length must be increased, leading to quadratic increases in costs. Attention demonstrates the bitter lesson: end-to-end learning beats architecture, but it also forces us to eat a bitter pill: a quadratic cost that we can’t scale forever. A transformer’s quadratic scaling hits a wall from 100k to 1M tokens. It’s no coincidence that this is precisely the range of maximum context lengths seen in current frontier models.

Comparing linear O(n), log-linear O(n log n), and quadratic O(n²) scaling vs. input sequence length (n). Top: Log-scaled y-axis shows relative growth. Bottom: Linear y-axis dramatically illustrates how quadratic costs dominate at scale.

Nature’s Solution

But the truth is, generality doesn’t require a quadratic cost. Look at a single cell. It has gigabases of DNA and a swarm of biomolecules, operating in a parallel soup with no centralizing control. By some definitions, each cell’s molecular interactions also fall in that threshold-circuit-like class: essentially parallel transformations, no large sequential pipeline. But cells offset this limitation with state—encoded in their components and their DNA, which holds a record of life’s entire lineage.

When cells need to store or recall specific bits of information, they literally write it into the genome. The canonical example is CRISPR. At its core, CRISPR functions as a chronological log of viral fragments—each one a DNA address that will guide CRISPR effectors like Cas9 or Cas12 to destroy an invading genome. These fragments represent viral infections that both the individual cell and its ancestral lineage have survived. The chronology is not just an artifact of the system but is essential. By remembering which viruses are more recent, the cell can invest more energy in making guided nucleases that target them than ones which occurred long ago.

CRISPR, better known for its use as a genome editing tool, is actually a temporal memory system.

The ARC AGI Prize nicely illustrates where today’s models stand. A single forward-pass language model—like GPT-4—gets about 0–5% accuracy on ARC tasks. Chain-of-thought–based systems that do a bit of test-time training get close to 50%. Meanwhile, a careful human solver will approach 100% if you just give them enough time to think.

Beyond Brute Force

Then, in late December 2024, along comes OpenAI’s o3, which nearly solved the entire benchmark. If you let the model “think less”—what they call high-efficiency mode—it answers 400 tasks at 82.8% accuracy for around $6,600. But if you let it think 172 times longer, exploring many more possible solution paths in parallel, it climbs to 91.5%. That final push costs an estimated $1.15 million (≈ 172 × $6.6k) in total compute. o3’s authors describe this not primarily as an engineering hack but as an ethical one, “deliberative alignment” wherein an overseer system that maintains the alignment of an kind of “forest of thought”, MCMC-sampled chains-of-thought which are kept in check. Their published numbers suggest that they pay a monstrous cost to keep all those partial reasoning traces in memory, to preserve consistency, and let the model think longer in pursuit of another apparent scaling law.

In effect, by brute-forcing memory into reliability, they show how easily you can approach universal computation if you’re willing to burn a fortune on tokens. And this isn’t like previous scaling laws that revealed new capabilities—it’s paying an enormous price to achieve something that biological systems do routinely. To reach the kind of sustained reliable thinking needed for real breakthroughs—novel cancer treatments, scientific discoveries, fundamental insights—we’d need to maintain computational state not just for a few more steps, but across weeks or months of focused investigation. Not just in one mind, but across communities of thinkers building on each other’s work. At quadratic cost, that’s not just unsustainable, it’s impossible.

Meanwhile, human civilization achieves this kind of extended cognition through more elegant means. Ideas evolve across conversations, papers, and experiments, maintaining state through distributed networks of sandwich-powered minds working in shared contexts.

Current AI research has thus discovered that “thinking longer” helps, but in their excitement over this empirical finding, many miss that reliable memory is the fundamental capability that makes computation universal. The field has embraced psychological concepts like Daniel Kahneman “System 1” (fast, intuitive) versus “System 2” (slow, deliberate) thinking to explain these improvements. o1/o3’s developers advocate dramatically scaling test-time compute to “hours, days, even weeks”, but I truly wonder how quadratic-memory transformer-based architectures will allow this to happen.

Soon after o3’s release, OpenAI’s Board said that to achieve AGI, we’ll just need (as paraphrased) “unimaginable sums of money”, proposing to restructure as a public benefit corporation to attract fresh capital. It turns out, they were talking about sums like the $500 billion dollars which will be invested over the coming four years in the Stargate AI infrastructure Project. It’s time to scale up, spin more GPUs, and chase universal intelligence through raw power.

o1 Forgets

If you’ve been playing with one of the o1-class models, you may have noticed a difficult pattern. While in a single response, they can produce entire, bug-free code with complex algorithms, when interacted with over a series of messages, they don’t behave like regular LLMs. They seem to forget past parts of the conversation, losing key nuance.

Some claim this is simply a feature—that you’re “using it wrong” if you try to chat with it, that you need to frontload all context into a single prompt. But this misses a fundamental point: OpenAI’s documentation explicitly states that o1 “discards reasoning tokens from its context” after each response. This isn’t a feature to adapt to—it’s an architectural flaw that violates a core requirement for universal computation: reliable access to the history of thought.

OpenAI’s visual explanation of how o1 manages its context.

While o1 can execute impressive one-shot computations within a single step, it cannot build reliably on its own reasoning across multiple exchanges. Each response starts fresh, without access to the deep cognitive trails that led to previous insights. This isn’t about prompt engineering or learning the right usage patterns—it’s a theoretical barrier to achieving true general intelligence. The model may achieve remarkable individual results through massive context windows and parallel search, but without the ability to maintain and build upon its own reasoning state, it remains fundamentally bounded.

A Path Forward

As a sandwich-eating universal intelligence, I’m not convinced. Humans got smarter not by growing our brains, but by distributing memory across society—spoken word, libraries, the internet—finding ways to preserve and share knowledge without a ruinous overhead. We’ve been doing it for a while:

Written memory persists: A papyrus from Herculaneum, 2000 years old, decoded by a space-time transformer model for the Vesuvius Challenge. As a child of Kentucky (where the project is organized), and a long time resident of Napoli (right around the bay from Ercolano scavi, I deeply ❤️ this.

What’s next? How do we skip the bitter pill and break through the scaling limits posed by transformers? I’ll bet on a resurgence of recurrent approaches that drastically reduce inference costs by maintaining a rolling hidden state, incrementally processing new information, thinking slowly and deeply when we need them to. Yes, we might need specialized modules for robust memory access—but if biology, evolution, and a good pen-and-paper can do it, so can we. Generality doesn’t require industrial-scale cooling towers and power plants. It just needs a reliable way to remember and recall what you’ve thought.

Let’s remember.