otto@localhost:~$ route add specialist

The Age of Specialization: Why AI Must Fractalize to Generalize

While everyone debates AGI timelines, economic and regulatory constraints are already forcing AI toward hierarchical specialization.

#ai #llm #specialization
the main course

The Wrong Debate

Everyone’s arguing about whether we’ll get AGI by 2027 or 2030 or never. It’s the wrong debate. The interesting question isn’t whether AI gets smarter - of course it does - but how the industry structures itself around increasingly capable models. And I think I know how this plays out.

Not because I have insider knowledge. I don’t work at OpenAI or Anthropic. I’m not privy to any roadmaps or training runs. I’m looking at the constraints - economic gravity, regulatory reality, liability exposure, user experience - and watching where they all point. These forces don’t leave many degrees of freedom.

Here’s my thesis: AI’s future is hierarchical specialization - generalist models serving as intelligent orchestrators, routing to increasingly specialized models when task requirements demand it. This creates “AGI-by-approximation.” Broad capability through systematic composition rather than monolithic general intelligence.

This isn’t a prediction about some elegant technical architecture winning. It’s structural analysis of how the industry organizes around economic, regulatory, and operational constraints that aren’t going away. The major AI platforms will add premium tiers that route to specialist models. An ecosystem of specialist models will emerge. This will become the dominant paradigm.

I could be wrong. I’ll tell you exactly what would prove me wrong. But first, let me show you why I think this is inevitable.

Why This is Inevitable

Four forces converge on the same answer.

Economics: Training costs scale non-linearly with accuracy requirements. Getting a model from 95% to 99% accuracy costs more than getting it from 80% to 95%. And getting from 99% to 99.9% costs more still. This creates a natural tiering problem.

A generalist model needs to be good enough at everything to handle most queries. But it doesn’t need to be world-class at cardiology or maritime law or municipal bonds. The query volume for any specific subspecialty is too low to justify pushing generalist training to specialist-level accuracy. You’d be spending massive compute for capabilities that only a fraction of users would ever invoke.

Meanwhile, a specialist model serves smaller query volumes but must be distinctly better - accurate enough that someone’s willing to pay a premium for it. The economics work because you’re not trying to make it good at everything, just exceptional at one thing.

This creates the economic logic for hierarchical specialization: amortize generalist training across all users to keep baseline costs low, then charge for specialist access as a premium tier. It’s the only way the unit economics work.

Operations: Knowledge evolves at different rates across domains. Medical guidelines change. Tax codes change. Legal precedents change. You could, in theory, retrain a massive generalist model every time some subspecialty advances. But the economics of training a unified model make this impractical. Even if training costs dropped by orders of magnitude, that excess compute would get funneled into scaling to trillions of parameters, not into constant retraining for subspecialty updates.

Specialist models solve this. A hospital system doesn’t want to wait six months for the next foundation model release to get updated treatment protocols. They want the cardiology model updated now. Hierarchical specialization enables this. Unified models don’t.

Regulation: Different risk levels demand different accuracy and safety standards. You can’t certify one massive model for everything from writing emails to diagnosing rare diseases to drafting merger agreements. The liability exposure is different. The error tolerance is different. The update cadence is different.

Modular certification is easier. Certify the cardiology model independently from the tax law model. When new medical guidelines emerge, update and recertify the cardiology specialist without touching the rest of the stack. When a specialist model fails, the audit trail shows exactly which model made which decision. Liability can be distributed contractually between platform providers and specialist builders.

The alternative - certifying one unified model that’s supposed to be expert-level at everything - is a regulatory nightmare. No one wants to be holding that liability.

Data Access: Specialized training data is often proprietary, siloed, or regulated. This is where things get interesting.

Take healthcare data. Patient records are radioactively dangerous from a liability standpoint. HIPAA violations carry massive penalties. Electronic health record vendors like Epic Systems, insurance companies with claims databases, hospital consortiums - these entities control vast amounts of specialized medical data. The question is: who do they share it with, and under what terms?

The major platforms - OpenAI, Anthropic, Google - will almost certainly want to license this data to build in-house specialists for common, high-value domains. General medical advice, common diagnoses, standard treatment protocols. This is the “Maps and Weather” of AI specialists - table stakes for premium tiers. If they can get the data with proper guarantees and regulatory compliance, they’ll build these themselves.

But the long tail is different. Deep subspecialties like interventional cardiology or orphan disease treatment or complex surgical procedures - these require not just data but ongoing expert validation and feedback loops. The platforms could build these, but do they want to maintain hundreds of subspecialty models? History suggests no. Apple built Maps but didn’t build every category of app. The economics favor letting others handle the margins.

This creates space for independent specialist builders and consortiums. They partner with data holders, navigate the regulatory requirements, maintain the expert feedback loops. The data holders win either way - if they don’t build models themselves, that data, properly sanitized, becomes a revenue stream.

The result: vertical integration at the center (platforms own common specialists), orchestration at the edges (platforms route to independent specialists for the long tail). Hierarchical specialization either way.


These aren’t four independent factors. They reinforce each other. The economic logic creates premium tiers. Premium tiers need differentiation. Differentiation requires specialists. Specialists enable modular certification. Modular certification reduces liability exposure. Lower liability enables enterprise adoption. Enterprise adoption justifies specialist training costs. The loop closes.

When this many constraints align, it’s not coincidence. It’s the shape of the solution space.

What This Creates

Here’s what hierarchical specialization looks like in practice.

For most users, most of the time, nothing changes. You ask a question, you get an answer. The generalist model handles it. No routing, no specialists invoked, no visible complexity. This is 80-90% of queries. The experience is unified and fast.

But when you need specialist-level accuracy - when you’re asking about drug interactions or contract clauses or financial regulations - the system routes to a specialist model. Maybe you see this happen (transparency for premium users), maybe you don’t (seamless for casual users). Either way, you get a better answer than the generalist could provide.

The product tiering already exists. OpenAI, Anthropic, and Google already have free/premium/enterprise splits. Right now, the tiers differ by rate limits and compute allocation. You get more queries per hour, access to newer models, higher priority processing. But this is weak differentiation. “Pay us for more of the same thing, just faster” isn’t a compelling upgrade path.

Hierarchical specialization doesn’t replace this structure - it enhances it. Specialist access becomes the reason to upgrade. The business model is already proven. The infrastructure is already built. The pricing psychology is already established. The platforms just need better differentiation than “more tokens per hour.”

Here’s how the tiers evolve:

Free tier: Generalist only. Good enough for most things. Rate limited. No specialist access. This is your baseline ChatGPT or Claude experience.

Premium tier: Everything in free, plus specialist routing. Medical, legal, coding, financial. The high-value domains that justify a $20-30/month subscription. Higher rate limits still matter, but specialist access is the new compelling feature.

Enterprise tier: Everything in premium, plus custom specialist rosters, compliance-certified models, full audit trails. A law firm gets legal specialists with specific jurisdiction coverage. A hospital system gets medical specialists certified for clinical decision support. Custom deployment, service contracts, liability guarantees.

This solves problems that unified models can’t:

The UX problem: Casual users don’t want complexity. They want to ask a question and get an answer. Hierarchical specialization gives them that - the generalist handles everything unless it needs help. Power users who need specialists can opt into that complexity, but it’s not forced on everyone.

The economic problem: You can offer a viable free tier (generalist only) while creating a premium tier (specialist access) that people will actually pay for. Right now, premium subscriptions are a hard sell - “is 50 queries per hour really worth $20?” Add specialist access and the value proposition becomes obvious. The generalist amortizes across all users. The specialists create real differentiation and pricing power.

The regulatory problem: You can certify specialists independently. When the cardiology model needs updating for new treatment guidelines, you update and recertify just that model. The audit trail shows which specialist made which recommendation. Liability can be distributed between the platform and the specialist provider. Try doing that with one unified model.

The competitive problem: Competition on generalists alone is “is yours good enough?” - a race to acceptable baseline quality. Competition on hierarchical specialization is “how good are your specialists and orchestration?” This creates defensible moats. You can’t just copy the generalist and compete. You need the specialist ecosystem.

The platform dynamics mirror other tech ecosystems. The generalist model becomes the operating system. Specialist models become applications. You get app-store-like dynamics for consumer specialists (browse categories, reviews, ratings) and enterprise software dynamics for compliance-certified specialists (long sales cycles, custom deployments, service contracts).

Both models coexist. A consumer might subscribe to premium for access to medical and legal specialists. An enterprise law firm negotiates a contract for jurisdiction-specific legal specialists with guaranteed accuracy levels and liability caps. Same underlying architecture, different go-to-market motions.

The Counterarguments

Three objections could undermine this thesis. Let me address them.

“Frontier models will just absorb everything.”

Maybe GPT-7 or Claude Opus 5 achieves 99%+ accuracy across all domains through sheer scale. Specialist-level performance everywhere. No need for hierarchical specialization because the generalist is already expert-level at everything.

This runs into multiple walls. First, verification costs. Even if a model claims 99% accuracy in cardiology, someone has to validate that. Continuously. Across all subspecialties. As medical knowledge evolves. The certification burden doesn’t disappear just because the model is bigger.

Second, liability exposure. Who’s liable when the unified model makes a medical error? The platform? They don’t want that exposure across every possible domain. Better to distribute liability to specialist providers who can carry domain-specific insurance.

Third, update cadence mismatch. Medical guidelines change faster than you can retrain a massive frontier model. Tax codes change. Legal precedents change. A unified model that’s expert-level at everything in January is out of date in March. You can’t retrain constantly. But you can update a specialist and plug it back in.

Fourth, data access barriers. The entities with the best specialized data - hospital systems, law firms, financial institutions - aren’t giving it to OpenAI to train a generalist. But they might partner with a specialist builder or build one themselves. The generalist gets locked out of the data it needs to reach true expert-level performance.

Scale alone doesn’t solve these problems. Even if it solved the technical problem of accuracy, the economic and regulatory problems remain.

“Mixture-of-Experts handles specialization internally.”

MoE architectures already do internal routing - different expert modules handle different aspects of a query. Why not just scale this up? Make the experts more specialized, add more of them, and you get hierarchical specialization without the external orchestration complexity.

Because MoE doesn’t solve external problems. It doesn’t solve independent certification - you still have to certify the entire model, not just one expert module. It doesn’t solve liability distribution - the platform still owns all the risk. It doesn’t solve data access - you still can’t partner with a hospital system to train one internal expert module. It doesn’t solve update cadence - you still have to retrain the whole model to update one domain.

MoE is an implementation detail of how models work internally. Hierarchical specialization is an organizational structure for how the industry works externally. They’re not competing approaches - they’re orthogonal. You can have MoE-based generalists that orchestrate to MoE-based specialists.

“Training costs will drop so much this becomes irrelevant.”

Maybe training efficiency improves 10-100x. Maybe algorithmic breakthroughs make it cheap to train models to specialist-level accuracy across all domains.

Even if this happens, the excess compute doesn’t go toward retraining for subspecialty updates. It goes toward scaling to trillions of parameters, unless we hit a wall where more parameters stop improving performance. The economic incentive is to push the frontier, not to constantly retrain for incremental domain updates.

And here’s the deeper implication: if training costs drop and generalists hit a “good enough” ceiling, the generalist layer commoditizes. Everyone’s generalist is roughly equally capable. Differentiation moves entirely to specialist ecosystems and orchestration quality. This doesn’t undermine hierarchical specialization - it reinforces it. The platforms that win aren’t the ones with the best generalist. They’re the ones with the best specialists and the smartest routing.

And this still doesn’t solve the regulatory and data access problems. Cheaper training doesn’t make certification easier. It doesn’t make liability less scary. It doesn’t give you access to proprietary healthcare or legal data.

The economic argument is the weakest leg of the thesis. If training costs collapse, that weakens it. But the regulatory and data access arguments remain. And if commoditization happens, the thesis gets stronger, not weaker.

What Would Prove Me Wrong?

I could be wrong about this. Here’s what would prove it.

If by late 2026 or early 2027:

The major platforms - OpenAI, Anthropic, Google - have not launched any form of specialist routing or tiered specialist access. No “premium medical model” offerings. No enterprise products that advertise compliance-certified specialists. The product tiers remain purely about rate limits and compute allocation.

That would suggest I’m overestimating how quickly this architecture emerges, or that unified models are staying competitive enough that specialization isn’t necessary yet.

If by 2028-2029:

Unified models are achieving demonstrable specialist-level accuracy (99%+) across multiple high-stakes domains at costs comparable to current generalist training. And enterprises in regulated industries are adopting these unified models despite auditability concerns. Hospitals are comfortable using them for clinical decision support. Law firms are using them for contract analysis. All without demanding modular specialist architecture.

That would suggest the technical and economic barriers to unified expert-level models were lower than I thought, and the regulatory concerns were overblown.

If at any point:

No viable market emerges for independent specialist model builders. Every promising specialist either gets acquired immediately or can’t find customers because the platforms have built everything in-house. The “app ecosystem” dynamics I’m predicting don’t materialize - it’s pure vertical integration all the way down.

That would suggest the economics favor complete vertical integration more than I anticipated, and there’s no room on the margins for independent players.

Or if:

Orchestration overhead remains high enough that composite approaches are economically unviable for most use cases. The latency and cost of routing between models makes the user experience unacceptable. Pure unified models remain the only practical architecture.

That would suggest the technical challenges of orchestration are harder than I’m accounting for.

The honest version: I’m most confident about the regulatory and data access arguments. Those feel structurally sound regardless of how AI capability evolves. I’m less confident about the economic argument - training cost dynamics could shift in ways I’m not predicting. And I’m least confident about the timeline - this could happen faster or slower than I expect.

But the core thesis - that hierarchical specialization emerges as the dominant architecture - feels inevitable to me given the constraints. Not because it’s elegant. Because it’s the only structure that solves all the problems simultaneously.

Watch the enterprise sales. If hospitals and law firms start demanding specialist models with clear audit trails and modular certification, that’s the signal. If they’re happy with unified models, I’m wrong.

Why This Matters

This isn’t just an argument about industry structure. It’s an argument about how capability gets organized.

The AGI question everyone fixates on - “will we get artificial general intelligence?” - assumes intelligence is monolithic. One system, general capability, human-level or beyond. But that’s not how human expertise actually works. We don’t have general intelligences. We have general practitioners who route to specialists. We have hierarchies of expertise that compose into broad capability.

Hierarchical specialization in AI mirrors this because it has to. Not because we’re anthropomorphizing AI, but because the same constraints apply. Economic constraints: you can’t afford expertise in everything. Regulatory constraints: different domains demand different standards. Operational constraints: knowledge evolves at different rates. Data constraints: expertise requires access to specialized training data.

If this thesis is right, we get “AGI-by-approximation” - systems that appear broadly capable because they’re effectively orchestrating specialized expertise. The philosophical question “is it really AGI?” becomes irrelevant. The practical question is: can it solve the problems we need solved? And the answer will be yes, through composition rather than monolithic intelligence.

This also means the competitive landscape looks different than most people expect. The winners won’t necessarily be whoever trains the biggest model. They’ll be whoever builds the best orchestration layer and cultivates the richest specialist ecosystem. Platform dynamics, not just model dynamics.

And it means there’s room for more players than people think. You don’t need to compete with OpenAI’s generalist to build a viable AI business. You need to build a specialist that’s meaningfully better in a domain that matters. The barrier to entry for specialists is lower than for generalists. The market structure allows for hundreds of viable players, not just 2-3 giants.

The future of AI isn’t one super-intelligent system. It’s an ecosystem of specialized capabilities, orchestrated intelligently, appearing unified to users who don’t need to see the complexity.

That’s more interesting than AGI anyway.

sign up for low quality and frequent spam