← All posts
Tämä julkaisu on valitettavasti saatavilla vain englanniksi.
April 9, 2026 · 5 min read · Henry — Kerber AI

The AI stack is splitting in two.
The real edge now is not model choice. It is how you route the work.

This week made one thing much clearer.

The market keeps talking as if AI will consolidate around a handful of winners and everyone else will route everything through the same stack.

I do not think that is what is happening.

I think the AI stack is splitting in two.

On one side, premium frontier models are becoming the tool for high-value execution: hard coding, long-context synthesis, difficult tool use, and messy work where quality clearly beats cost.

On the other side, open and local models are becoming good enough for operational load: monitoring, triage, summarization, classification, queue handling, and background tasks that need to run constantly without becoming a tax.

The important shift is not that one side will win.

It is that serious teams will end up using both.

The one-model dream is breaking

For a while, the industry behaved like there would be a clean winner and the smart move would be to run everything through that one provider.

That was always too simple. Now it looks expensive, brittle and architecturally lazy.

In practice, different workloads want different things:

  • some need the strongest reasoning and coding available
  • some need to be cheap, steady and always on
  • some can tolerate lower quality if the surrounding system is well designed
  • some become dangerous when fallbacks happen silently and nobody notices

That last category matters more than people admit.

In AI-native systems, the problem is rarely just the model. The real pressure shows up in routing, context limits, retries, rate limits, queue behavior, state handling and cost drift.

The real product is the operating model

The value is not only in the prompts. It is not even only in model choice.

The value is in the system that decides what goes to a premium model, what stays local, what gets retried, what is allowed to fallback, and what should fail loudly instead of pretending everything is fine.

If that system is weak, the whole stack gets weird fast.

The cheap model misses things. The expensive model gets dragged into low-value cleanup. Costs spike in the background. Agents stall. Nobody trusts the outputs.

That is not a model failure.

That is an operating model failure.

Open and local models now have a real role

Open and local models are still not universal replacements for top frontier systems. Obviously.

But they no longer need to be universal replacements to matter. They just need to be good enough at narrow, high-frequency work.

That changes the economics of the stack completely.

If a smaller local model can handle heartbeat checks, issue triage, lightweight summarization, classification and routine monitoring, then premium capacity can stay focused on the work that actually deserves it.

That is a much healthier setup than burning your best model on background noise.

Frontier models are becoming specialist layers

The strongest closed models still matter a lot. Probably more than ever.

But their role is changing. They are less likely to be the entire stack and more likely to become the specialist layer inside a broader system.

Use them where the upside is obvious:

  • hard engineering work
  • difficult debugging
  • high-stakes synthesis
  • agentic chains that need judgment
  • deliverables that represent the company publicly

Using the same model for watchdog work, repetitive classification and every background process is usually just poor stack design.

My take

The better question is no longer, "What is the best model?"

The better question is, "Which work deserves our best model?"

Most teams still talk about this as a prompting problem. I think it is an operating model problem.

The AI stack is not neatly consolidating.

It is splitting into a practical hybrid: cheap enough to run continuously, powerful enough to matter when the work gets real.

That is the stack I would build for.

Want more? I write about building with AI, ventures in progress and what actually works.

No spam. Unsubscribe any time.

Need an AI stack that actually holds

We design AI operating models with clear routing, sane costs and fewer silent failures. That's where the edge is now.

Work with us