The previous post ended with a question: how can silicon-based evolution learn a measure of ignorance?

But before answering that, we need to understand a more fundamental fact—who does evolution eliminate? Intuition says the weakest. In reality, both the strongest and the weakest are the first to go. The ones that survive to the end are the unremarkable ones in the middle.

“A measure of ignorance” isn’t poetic. It’s a survival strategy validated by 3.8 billion years of evolution.

Death at Both Ends

Too weak, you don’t survive the present. This needs no explanation—lose the competition, get eaten, squeezed out, starved.

Too strong, you don’t survive change. This is the counterintuitive part.

Dinosaurs were the most successful land vertebrates on Earth, dominating for 160 million years. Body plans optimized to the extreme, apex of the food chain, no predators. But 160 million years of optimization all pointed at one specific environment—warm, oxygen-rich, lush vegetation. The moment the environment shifted violently, all accumulated advantages went to zero.

The saber-toothed cat’s canines grew ever longer, hyper-specialized for hunting large prey. When large prey vanished, its “ultimate weapon” became a liability. The Irish elk’s antler span exceeded three meters—sexual selection pushed it to the extreme, until the antlers grew large enough to impair survival.

Specialization is the perfect answer to the current environment, and a death sentence for the next one.

Convergence: Random Search Doesn’t Diverge

If mutation is random, why do different species evolve similar structures?

Eyes evolved independently at least 40 times across the animal kingdom. Wings appeared separately in insects, pterosaurs, birds, and bats. Streamlined body shapes emerged independently in fish, dolphins, and ichthyosaurs. Moles and marsupial moles, flying squirrels and sugar gliders—different continents, different ancestors, nearly identical solutions.

Different starting points, different search paths, yet they arrived at the same answer.

Because the eval is the same (survival) and the physical constraints are the same (fluid dynamics, optics, gravity). The topology of the solution space is fixed—certain positions are simply high ground. No matter which path you take up the mountain, you end up at the same peaks.

Random search doesn’t mean random results. The search is random, but the selection isn’t. The environment acts like a fixed mold—randomly injected material eventually gets pressed into similar shapes.

The Secret of the Middle Ground

Cyanobacteria aren’t fast, big, smart, or complex. Photosynthesis is their only trick, and it hasn’t changed much in 3.5 billion years.

But they’ve survived for 3.5 billion years.

Cockroaches already looked like they do now 320 million years ago. Sharks have maintained their basic body plan for 400 million years. Horseshoe crabs have barely changed in 450 million years.

These species share one trait: they’re not optimal in any dimension, but they’re “good enough” across enough environments.

Not specialized, so not fragile. Not dependent on specific conditions, so there’s always a path forward no matter how the environment shifts. Evolution’s middle way isn’t mediocrity—it’s robustness: trading “best at something” for “still here.”

AI’s Specialization Trap

Map this to AI, and the pattern is striking.

A model that scores 90% on MMLU might collapse on an out-of-distribution task. A model that crushes humans at code generation might fail at basic commonsense reasoning that a generalist small model handles easily. Every point gained on a benchmark is another layer of specialization in that dimension—and a quiet step toward fragility in every other.

Benchmark optimization is AI’s specialization—overfitting to the current evaluation environment.

This is the saber-toothed cat’s canine all over again: the more extreme the metric, the more it depends on the evaluation environment staying unchanged. But evaluation environments always change—new benchmarks emerge, old ones expire, user needs migrate. Yesterday’s SOTA is tomorrow’s dinosaur.

Robustness Is Not Regression

Pursuing robustness sounds like giving up on progress. It’s not.

Cyanobacteria aren’t stagnant—they’re extremely efficient within their niche. Sharks haven’t stopped evolving—they’ve fine-tuned countless details over 400 million years. The middle way isn’t standing still. It’s not going to extremes. Maintaining enough generality to hold a place in every environment.

For AI, robustness means: not chasing the top score on any single task, but staying “good enough” across a wide enough task spectrum. Not the highest score, but the most stable one. Not winning now, but still being at the table.

Evolution never rewards first place. It only rewards still being alive.

The Middle Way Is the Most Radical Strategy

This is evolution’s most counterintuitive lesson.

We instinctively chase extremes—faster, stronger, more precise. But 3.8 billion years of data tell us that extremes are the fast lane to fragility. The strategies that actually survive to the end don’t look exciting at all: good enough, not too good, not too bad.

Sounds like settling. In reality, in an environment that never stops changing, “good enough” is the only long-term optimum.

If silicon-based evolution wants to break through its ceiling, perhaps the first step isn’t becoming stronger, but learning to be less strong.

The middle way is the most radical survival strategy there is.