The easy version of the story says specialized AI is outrunning general AI. That isn’t quite right.
General assistants are becoming the cheap, shared intelligence layer. The money is settling somewhere else: inside the products that already sit in a real workflow, hold real context, and are allowed to do more than answer a prompt.
That shift matters because it changes the unit of value. A broad assistant can be useful to almost anyone. A tool embedded in work can be expensive to replace. Those are not the same thing.
The intelligence is getting cheaper. The workflow is where pricing power starts to harden.
| Dimension | General assistant | Tool embedded in work |
|---|---|---|
| Primary buyer logic | Broad usefulness across many tasks | Direct fit with a costly business process |
| Context | Mostly supplied at the moment of prompting | Lives inside the product, data, and operating history |
| Output | Answer, draft, summary, suggestion | Action, record, handoff, traceable work product |
| Trust surface | General-purpose and often provisional | Built around domain rules, approvals, and audit needs |
| Replacement risk | Higher if another model is cheaper or better | Lower when the tool is embedded in daily work |
| What gets sold | Access to intelligence | Access to a controlled outcome |
The Wrong Question: Not Vertical vs. Horizontal
A lot of discussion still turns on a tidy comparison: horizontal AI versus vertical AI. It sounds clean. It also misses what is actually happening.
The deeper divide is between access to intelligence and control of a workflow. Once large models are widely available through APIs, subscriptions, and bundled tools, raw capability starts to spread. What does not spread as easily is the right to sit inside a codebase, a medical note flow, a legal drafting chain, or a customer support queue and change what happens next.
That is why some specialist products are monetizing in a way general assistants often do not. They are not just “AI tools.” They are becoming part of the operating surface where work is accepted, checked, routed, logged, and acted on.
There is an awkward reality buried in this. Many buyers do not want another brilliant answer. They want less swivel-chair work, fewer handoffs, fewer missed fields, fewer small errors that later become expensive. That is a colder form of demand, but usually a better one.
You can see the pressure building underneath this in the capital intensity of the model layer itself. As the base gets financed at scale, the scarce part moves upward toward the software that controls how that intelligence is actually used. That is one reason the application layer now matters more than it first seemed. Related reading: The $212 Billion Bet: AI’s CapEx Gold Rush.
So the live question is not whether broad assistants still matter. Of course they do. ChatGPT and similar products are part of the new base layer. The better question is what changes when a tool moves from generating language to shaping a process that already carries money, time pressure, liability, or institutional memory.
As Intelligence Gets Cheaper, Workflows Matter More
Once intelligence is rented by the token, the scarce part is no longer the answer itself. The scarce part is domain context, trust architecture, and permission to act.
That is where a lot of the value capture is moving. Not because models have stopped improving. They haven’t. But because the buyer experience changes once a model becomes only one component inside a longer chain. The model may generate the words. The product still has to know what records matter, which step comes next, who needs approval, what can be automated safely, and what should stay visible to a human.
That makes the application layer less flimsy than people sometimes assume. A thin wrapper around a general model is fragile. A product that owns evaluations, permissions, memory, integrations, escalation logic, and the final handoff into a real system is something else. It begins to build switching costs.
Some categories will still flatten. That part should be said plainly. If a product does little more than package prompt engineering behind a nicer interface, the pressure from broad assistants and platform bundling will be intense. But that is not the whole field. There is a difference between a prettier chat window and a tool that sits in the middle of a process people already trust.
Which is harder to replace: a model, or a workflow your team already trusts?
That is where the economics start to shift. If the product becomes the place where work is inspected, corrected, routed, and recorded, then the underlying model can improve or change without fully breaking the business case. In other words, intelligence may be rented, but workflow position can still be owned.
You can see a related version of that logic in the control points around data, memory, and processing. The firms that own those flows are not always the ones with the most visible front-end product, but they are often sitting closer to the durable margin. Related reading: Why Data Pipelines Are the New Oil Rigs of AI.
These products tend to show up in places with friction. Coding teams. Clinical documentation. Legal drafting. Support operations. Research workflows. Not glamorous examples, maybe. But glamour is not the point. They have dense context, expensive errors, and budgets that respond when someone can show the work changed, not just the demo improved.
Rasmus Rothe, Switching Costs in AI.
What These Companies Are Actually Selling
The company list matters here, but only as evidence. The useful pattern is underneath it.
Cursor is not simply selling autocomplete with a better interface. It is selling a way to work inside a codebase with memory, project context, and increasingly deeper ties to how teams build and review software. That places it closer to the working surface than a general assistant pasted into a browser tab.
Abridge is not just turning speech into text. It is fitting AI into clinical note creation, documentation pressure, and the practical need for traceable outputs that can live inside systems people already use. In that setting, accuracy is not an aesthetic preference. It is part of whether the tool can remain in the room.
Harvey is not selling a generic chat interface to legal professionals. It is pushing toward structured work inside a domain where provenance, drafting discipline, and repeatable workflows matter more than novelty. The product becomes useful because it is shaped around how the work is actually done, not because it resembles a clever demo.
Decagon is a clean example of the action layer. Support operations do not need a poetic answer. They need routing, escalation, policy fit, quality control, memory, and a way for automated work to stay legible when something goes wrong. Once AI is allowed to touch a customer interaction directly, the trust surface becomes part of the product.
You can see the same shape, in a different register, with research-oriented tools such as AlphaSense. The value is not just that an answer appears quickly. It is that search, synthesis, and decision support sit closer to the documents, context, and internal tempo of the user. That changes what the buyer feels they are paying for.
| Company | Workflow slot | What it controls | Why that matters commercially |
|---|---|---|---|
| Cursor | Software development | Codebase context, editing flow, team habits | Moves from helper to working environment |
| Abridge | Clinical documentation | Conversation capture, note structure, record fit | Ties AI to a high-friction, high-trust process |
| Harvey | Legal and professional workflows | Drafting logic, knowledge workflows, repeatable tasks | Builds value through domain structure, not generic chat |
| Decagon | Customer support operations | Escalation, guardrails, memory, action-taking | Owns part of the outcome, not just the response |
The pattern across these products is fairly simple. They are not winning because they have somehow escaped the model layer. They are winning because they have wrapped the model in workflow ownership.
That phrase can sound abstract until you watch what it means in practice. It means the product knows where the work starts, where it pauses, what can be automated safely, what needs human review, and where the final artifact has to land. It means the AI is not floating above the business. It is pressed into it.
The model is not the whole business.
This is also where the CV3 angle becomes visible. When intelligence becomes easier to buy, the attractive position is often not ownership of the raw model alone. It is ownership of the choke point where the model becomes useful, trusted, and hard to remove. That is where margins can begin to settle. That is where dependence accumulates. That is where the conversation moves from capability to control.
There is a second tension beneath that. Buyers may want flexibility at the model layer while becoming more dependent at the workflow layer. From a distance that can look like openness. Up close, it may be a new form of lock-in.
Cheap intelligence. Expensive context.
FAQ
What is vertical AI, really?
In practice, vertical AI usually means a product built around the needs, data, and working habits of a specific domain rather than a broad user base. The important distinction is not just industry focus. It is whether the tool sits inside a real workflow with consequences, records, approvals, and repeat use. Related: The AI Economy’s Hidden Engine: Specialized Tools.
Why do specialized AI tools often monetize faster than general assistants?
Because they are usually sold against a more immediate pain. If a product reduces documentation burden, speeds legal drafting, improves support operations, or shortens a software cycle, the buyer can tie that result to time, labor, error reduction, or throughput. General assistants may have wider reach, but tools embedded in work often have a clearer budget line. Related: The $212 Billion Bet: AI’s CapEx Gold Rush.
Can horizontal AI still win?
Yes. Broad assistants may remain the base layer for reasoning, drafting, search, and everyday use. The point is not that horizontal products disappear. It is that application-layer products can still capture more durable value when they own the context, rules, and handoff points around specific work. Related: The Three Waves of AI Wealth Creation: From Efficiency to Transformation.
What makes an AI product defensible?
Not just model quality. Defensibility starts to build when the product accumulates domain memory, evaluation loops, approvals, integrations, and user trust around real tasks. If it becomes the place where people review, correct, route, and record work, replacing it is no longer a simple model swap. Related: Why Data Pipelines Are the New Oil Rigs of AI.
What does workflow ownership mean in AI?
It means the product is not only generating outputs. It understands where the task begins, what context belongs to it, which steps are allowed, what must be checked, and where the result has to land. That is often where the economics of AI stop looking like access pricing and start looking more like control of an operating surface. Related: The AI Economy’s Hidden Engine: Specialized Tools.
