The Cannibalization Trap

The Cannibalization Trap

Why seat-based pricing relies on a correlation that no longer exists.

In the SaaS era, seat-based pricing wasn't just a business model but a proxy for value to both the buyer and the seller. A proxy that made sense when human headcount was the unit of work. A proxy that breaks completely when AI is.

SaaS for Compliance

We touched upon the constraints in the SaaS era and how progress was measured as a function of scale here.

When a VP of Sales buys 100 seats of a CRM, they aren't paying for 100 people to sell better. They are paying for a standardized process - to measure an account going south, to create a digital trail for the post-mortem, and to get dashboards to monitor progress.

In other words, products supported the organizational structure that facilitated compliance.

Look at the titans of the SaaS era - Salesforce, Workday, ServiceNow. These platforms function as Systems of Record and provide the organizational hygiene that allows leadership to feel a sense of control.

This worked beautifully for vendors because it aligned their revenue with the customer's inefficiency. If a customer's process was bloated and required 50 people to manage a workflow, the vendor got paid for 50 seats. The vendor had zero incentive to automate the work, because automating the human out of the loop meant automating the revenue off the balance sheet.

SaaS companies became Digital Landlords. They rented you the desks (seats), but they didn't care if work got done.

The SaaS landlord model: revenue aligned to inefficiency
Seat-Based Pricing as a Proxy
A beautiful alignment — for the wrong thing
The Buyer
Paying for
Process Control
"We bought 100 seats of the CRM." — Not to sell better, but to standardize the process, create a digital trail, and get dashboards for leadership.
Buying process, not outcomes — dashboards, audit trails, compliance
"We bought the tool. Adoption is high. CS team is keeping us warm."
Seat count signals investment, not productivity
Never had to define what success actually looked like
The Vendor
Charging for
Inefficiency
If automating the customer saves 40 seats, the vendor loses 40% revenue. They delivered value — and their reward was a revenue cut.
Revenue aligned to customer's inefficiency, not their results
Zero incentive to automate — automation reduces seat count
The more bloated the customer's process, the higher the ARR
Digital Landlord: rented the desks, didn't care if work got done
SaaS companies became Digital Landlords. They rented you the desks (seats) — but didn't care if work got done. Revenue grew with your headcount, not your outcomes.

The Decoupling of Reasoning

In the Agentic AI era, AI agents are active participants in your workflow - moving from a System of Record to a System of Action (powered by the Context Graph, see The Agentic Transition) and breaking all the proxies.

Reasoning and actions are now decoupled from headcount scaling.

This creates a Cannibalization Trap for legacy vendors:

  • Scenario: A law firm uses an AI Agent to draft contracts.
  • The Impact: The AI allows 1 senior lawyer to do the work of 5 junior associates.
  • The Trap: If the vendor charges by the seat, they just helped the customer fire 4 users. They successfully delivered value, and their reward was an 80% revenue cut.
The cannibalization trap: deliver value, lose revenue
The Law Firm Scenario
Seat-based pricing meets AI efficiency
Before AI
5 Junior Associates
Annual seat revenue
$120k
5 seats × $24k/year
After AI Agent
Click to deploy →
Annual seat revenue
Awaiting deployment
AI allows 1 senior lawyer to do the work of 5 junior associates.

Building the Walled Gardens

Legacy vendors face the Innovator's Dilemma as AI agents penetrate into workflows. And they cannot afford to let external, efficient agents cannibalize their seat count.

They will build walls and block external agents, create exclusivity for their own proprietary agents, and force you to consume AI through their existing, seat-heavy interfaces and business models.

They are trying to extend their relevance by controlling the efficiency. They are fighting to keep the Input Metric (Seats) alive in an Outcome World. It will work for a while - until it doesn't.

The walled garden playbook: fighting to keep the input metric alive
The Innovator's Dilemma Response
Legacy vendors cannot cannibalize themselves
01
Tactic 01
Block external agents from accessing data
API restrictions, proprietary data formats, and integration walls that prevent efficient outside agents from touching the workflow.
Shelf life
2–3 years
02
Tactic 02
Create exclusivity for their own proprietary agents
Force you to buy their AI add-on at a premium, bundled with the existing seat contract — AI as upsell, not replacement.
Shelf life
3–4 years
03
Tactic 03
Maintain seat-heavy interfaces for AI consumption
Design AI features that require human interaction at every step — ensuring the seat count stays intact despite the automation.
Shelf life
2–3 years
04
Tactic 04
Lobby for regulatory complexity around AI in enterprise
Use compliance requirements and audit trails as a reason to maintain human-in-the-loop at every node — even when unnecessary.
Shelf life
4–5 years
They are fighting to keep the Input Metric (Seats) alive in an Outcome World. It will work for a while — until it doesn't. The market doesn't ask permission when it shifts.

The Definition Problem

So, why haven't we switched to Outcome-Based Pricing yet?

Outcome-Based Pricing forces a conversation that most enterprises desperately try to avoid: What exactly is the outcome?

Seat-based pricing allowed organizations to be lax. "We bought the tool, adoption is high, the CS team is keeping us warm." - was a fluffy, but safe and compliant response so far.

But outcome pricing removes the fluff with Radical Clarity. It's less like a Gym Membership (paying for access regardless of results) and more like Medical Treatment (monitoring vitals: did the doctor diagnose the issue? Is the medication working? Am I getting better?).

Gym membership vs. medical treatment: the accountability shift
Seat-based pricing
Gym Membership
Paying for access, regardless of results.
Model
Access-based — pay regardless of results
Success
"Adoption is high. CS team is warm."
Accountability
Buyer: "We bought the tool."
Failure mode
Gym profits when you stop going
Clarity
Fluffy — safe and compliant
Outcome-based pricing
Medical Treatment
Monitoring vitals. Paying for getting better.
Model
Outcome-based — pay when vitals improve
Success
Did the problem get solved? Measurably.
Accountability
Buyer must define the spec. Seller guarantees.
Failure mode
Doctor gets fired if the patient doesn't improve
Clarity
Radical — uncomfortable but honest

If you don't see progress, you stop paying. That level of accountability scares both the buyer (who has to define the spec) and the seller (who has to guarantee the result). This requires hardcore change management, and companies only do this when an external stimulus (like when the market shifts under their feet) forces them to.

Challenge of Attribution

We won't jump straight to revenue sharing. That's a leap of faith for anyone, and so it'll likely be a phased journey.

Level 1: Verifiable Cost Savings (The Defensive Moat)

This is already happening today. "Our Agent deflected 4,000 calls. At $5 per call, that is $20,000 saved." - easy to measure, easy to verify, and fits into existing budget lines.

Level 2: Top-Line Attribution (The Ultimate Moat)

It isn't just cost savings or resource augmentation, but tapping into the quality of the reasoning by the AI agents in creating the desired outcomes.

The attribution dashboard: from cost center to revenue partner
Agentic Attribution
The two levels of proof
Level 1 — The defensive moat
Verifiable Cost Savings
"Our Agent deflected 4,000 calls. At $5/call, that's $20,000 saved."
Level 2 — The ultimate moat
Top-Line Attribution
Revenue influenced, broken down by autonomy level and human intervention rate.
Total revenue influenced
$500,000
Fully Autonomous — no human intervention
45%
Human-Assisted — 1–2 corrections
20%
Human-Guided — reasoning updated
25%
Human-Overridden — AI was wrong
5%
AI reasoning was 95% accurate in enabling outcomes — directly contributing to the bottom line. You are no longer a cost center. You are a partner actively coexisting with humans.
The ultimate litmus test for any AI vendor: "If your AI makes me twice as efficient — does your revenue go up or down?"

Imagine a dashboard like this:

  • Total Revenue Influenced: $500,000
  • Fully Autonomous: 45% of deals (No human intervention).
  • Human-Assisted: 20% of deals (1-2 corrections).
  • Human-Guided: 25% of deals (Reasoning updated via feedback loop).
  • Human-Overridden: 5% of deals (AI was wrong).

When you can present this kind of data, you are no longer a cost center. You are a partner actively coexisting with humans. You have demonstrated AI's reasoning was 95% accurate in enabling the outcomes and directly contributed to the bottom line.

Conclusion: Agentic Economics

The SaaS era was built on Distribution Moats and Seat-Based Rents.
The Agentic era will be built on Attribution Moats and Outcome-Based Value.

For the buyer, this is the ultimate litmus test for any AI vendor:
"If your AI makes me twice as efficient, does their revenue go up or down?"

There are no easy answers here. Legacy vendors must cannibalize their own business models or be drifted out by shifting market forces. But the truth is, the Agentic market is far bigger than the SaaS market. It captures the Work, not just the Tool (but only if you have the Trust Budget to deploy it).

Agentic AI demands Agentic Economics.
If the software does the work, it must charge for the work.