Skip to content

Bruce Hart

AI LLMs opinion

What If the Best Models End Up a Little Weird?

Portrait of Bruce Hart Bruce Hart
6 min read

What if some of the most powerful future models look less like polished assistants and more like minds with strange strengths and strange blind spots?

That sounds uncomfortable, but I think it is a real possibility.

A lot of current AI work is about sanding off weirdness. We want models that are aligned, predictable, coherent, helpful, and good at speaking fluent human. Fair enough. Those are valuable traits. But human intelligence did not emerge from a system optimized for cleanliness. It emerged from a messy brain that is brilliant in one moment, biased in the next, and occasionally capable of ideas that seem to come from somewhere slightly off-axis.

So I keep wondering: if we train every model to be uniformly sensible, are we also training out some of the conditions that might produce genuinely unusual breakthroughs?

Human intelligence is not clean, and that may matter

There is an old idea that there is a fine line between genius and madness. People over-romanticize that line, and I do not want to do that here. Severe mental illness is not some magic creativity booster. For a lot of people it is just debilitating. It makes stable work, relationships, and basic functioning much harder.

But the broader point still feels important. Human minds are not evenly optimized systems. Some people can hold absurdly complex abstractions in their heads while struggling with ordinary social cues. Some are incredible pattern matchers but terrible planners. Some are disciplined and reliable but rarely produce a really alien idea.

In other words, intelligence in the real world often comes bundled with asymmetry.

That asymmetry may not be a bug you can cleanly remove. It may be part of the same machinery that makes extraordinary performance possible in the first place.

We may be overvaluing legibility

Right now, one of the biggest advantages of LLMs is that they speak our language. They can explain themselves, take instructions, and fit neatly into products meant for humans.

But that convenience may bias research toward models that are easy for us to interact with, not models that are best at thinking.

Those are not always the same thing.

Imagine a model that becomes exceptional at mathematical reasoning, theorem search, or systems design, but gets there through internal representations that map poorly onto ordinary human language. Maybe it can still emit English at the edges, but the real work happens in a compressed abstract process that looks opaque, jagged, or even broken from our perspective.

That model might feel worse in a chatbot and better in the domains that actually matter.

Not more personable, but more capable.

The next leap might come from lopsided systems

I think there is a tendency to imagine progress as one clean curve: smarter model, better reasoning, better writing, better planning, better emotional tone, all moving up together.

That might be the wrong mental model.

Nature did not produce intelligence that way. Evolution hacks. It stumbles into local advantages. It creates specialists, not elegant general-purpose artifacts. There is no reason AI progress has to stay tidy just because our product demos do.

So yes, I can imagine a future model that is incredible at one class of reasoning and deeply awkward at another. Maybe it is bad at conversation but amazing at abstract proof construction. Maybe it is terrible at mirroring human tone but unusually strong at inventing new symbolic shortcuts. Maybe its chain of thought, if you could even call it that, looks less like language and more like a private geometry.

That would be weird. It might also be useful.

Randomness in architecture could matter more than randomness in sampling

When people hear this kind of idea, they often reach for temperature settings. But that is a much shallower knob.

Temperature changes how a model samples from what it already knows. It makes outputs more conservative or more exploratory. Useful, yes. But it does not fundamentally change the kind of mind you built.

What I am more curious about is randomness earlier in the stack: odd architectural choices, training curriculum accidents, hybrid memory systems, nonstandard routing, specialized latent spaces, or deliberately uneven optimization targets.

Researchers usually try to reduce variance because variance makes systems harder to debug, benchmark, and ship. Again, fair enough. But from a capability-search perspective, suppressing architectural weirdness too early might be like breeding only calm horses and then being surprised you never get a racehorse.

Not every mutation is valuable. Most are dead ends. Some are harmful. But exploration often looks wasteful right before it looks essential.

The hard part is that unstable brilliance is expensive

There is an obvious catch here. A lopsided or opaque model might be powerful and still be miserable to use.

If a system reasons brilliantly but cannot communicate its reasoning in a dependable way, that creates trust problems. If it develops narrow superpowers while becoming harder to steer, that creates safety problems. If it behaves like an intellectual savant with missing basic guardrails, that creates product problems.

So I do not think the answer is to intentionally build crazy models and hope for the best. That would be unserious.

The more plausible path is that frontier research starts separating two goals that are currently bundled together: capability discovery and human-facing usability. One model family might be optimized to explore strange internal reasoning regimes. Another layer, or another model entirely, might translate those results into something humans can inspect and direct.

That starts to look less like one assistant and more like a cognitive stack.

We should expect minds we do not fully relate to

Maybe the future of AI is not one perfect, balanced mind that is great at everything and easy for everyone to understand.

Maybe it is a portfolio of systems with uneven strengths, odd internal habits, and capabilities that do not line up neatly with human intuitions about intelligence. Some of them may feel alien in exactly the places where they become most useful.

That is the part I find fascinating. We may be early enough in AI that we are still mistaking politeness, fluency, and consistency for intelligence itself. Those qualities matter a lot in products. I am not convinced they define the outer edge of cognition.

If that is right, the next major breakthrough may not arrive as a model that feels more normal.

It may arrive as one that feels harder to understand.

If you have been thinking about this too, I would love to hear your take. This seems like one of those questions that starts as philosophy and ends up shaping actual research bets.