Skip to content

The Bundled AI Bet: Why Enterprises Are Losing a Race They Think They're Running

2026-04-16 · 9 min read

Last week I wrote about corporations cutting headcount on the promise of AI that isn't ready yet — Oracle's 30,000 layoffs against record profits being the most visible example. This piece is the other side of that story: the AI that is ready, and the growing evidence that most enterprises aren't using it.

The question isn't whether organisations are adopting AI. They are — at significant scale and speed. The question is which AI, chosen how, and what the compounding cost of that choice looks like twelve months from now.

We Have Seen This Before

In the early 2000s, Siebel Systems was the dominant enterprise CRM vendor. Deeply integrated into existing IT infrastructure, trusted by procurement teams, already in the budget cycle. The objections to Salesforce were familiar: not enterprise-grade, security concerns, data governance questions, where does our data actually live?

Those objections were not wrong. They were just temporary. Salesforce closed the enterprise readiness gap faster than Siebel upgraded its capability. By the time the integration gap had closed, the capability differential was impossible to justify. Siebel is now a footnote.

Historical Parallel — The Bundled vs Best-of-Breed Pattern
Then — CRM (2000–2010)
Siebel: deeply embedded, trusted by IT, already in the stack
Salesforce: capability-first, initially dismissed as "not enterprise-grade"
Salesforce closed the integration gap faster than Siebel upgraded capability
Enterprises ran hybrid briefly, then the capability differential became unjustifiable
Now — AI (2023–?)
Copilot: deeply embedded in M365, trusted by IT, already in the budget
Claude, ChatGPT: capability-first, initially questioned on enterprise readiness
AI-first tools are closing the integration gap quarter by quarter
Enterprises are running hybrid now — the capability differential is already visible
The objections to AI-first tools in 2023 — security, governance, data residency — were legitimate. By 2026, enterprise tiers with SOC 2, SSO, data isolation, and audit logging are standard across Claude, ChatGPT, and Gemini. The 'not enterprise-grade' objection has expired. The question now is purely capability.

This is not a prediction. It is a pattern recognition. The bundled vs best-of-breed dynamic has played out in CRM, in productivity software, in ERP, in analytics. The timeline varies. The direction does not.

The Current Landscape

The "enterprise vs frontier AI" framing that appears in much current analysis is imprecise in a way that misleads. Claude, ChatGPT, and Gemini are all enterprise-grade in 2026 — they have been for some time. The meaningful distinction is between three categories with genuinely different value propositions:

PlayerCore value propositionIntegration trajectory
Microsoft Copilot
Bundled AI
Ecosystem integration. AI is the feature; M365 is the product. Value comes from proximity to existing workflows — not from raw model capability.
Deepening within Microsoft stack. Integration ceiling is high; capability investment trails AI-first competitors on independent benchmarks.
Claude, ChatGPT
Best-of-breed AI
Capability is the product. Enterprise features — compliance, SSO, governance, audit — added on top of best-in-class models. Enterprise-grade since 2024.
Aggressively closing the integration gap. Native connectors, API ecosystems, M365 and Google Workspace plugins. The moat is narrowing every quarter.
Gemini
Hybrid
Attempting both simultaneously: frontier capability AND deep Google Workspace integration. The most ambitious strategic position — and the most complex to execute.
Native to Google Workspace, expanding beyond. Unique long-context capabilities (1M+ token window). The most interesting competitive case to watch.

The strategic risk for enterprises is not that they chose a tool with strong integration. It is that they chose a bundled tool at a moment when best-of-breed tools are aggressively closing the integration gap — while the capability gap between those tools continues to widen in the opposite direction.

The Capability Gap Is Real and Measurable

The clearest independent measure of AI capability over time is the Chatbot Arena leaderboard — a crowdsourced benchmark where models compete head-to-head in blind pairwise evaluations across millions of user votes. Unlike static benchmarks, it measures what users actually prefer across diverse real-world tasks. It is imperfect, but it is the most defensible proxy available for output quality at scale.

Thirty-six months of Arena data produce a clear picture.

Figure 1
Chatbot Arena Elo — 36-Month Trajectory
Best-of-breed model capability vs Microsoft's Arena presence, May 2023 – April 2026. Microsoft held #1 for 2 of 36 months (WizardLM-70B, late 2023). Copilot has not appeared in Arena top rankings.
Best-of-breed #1Microsoft best
Q2 '23
1,148
1,098
Q3 '23
1,153
1,110
Q4 '23
1,207
1,207
Q1 '24
1,247
1,155
Q2 '24
1,260
1,140
Q3 '24
1,314
1,138
Q4 '24
1,365
1,135
Q1 '25
1,380
1,130
Q2 '25
1,450
1,128
Q3 '25
1,468
1,125
Q4 '25
1,490
1,122
Q1 '26
1,500
1,120
Source: BenchLM.ai Arena Elo Tracker, LMSYS Chatbot Arena dataset. April 2026 leaders: Claude Opus 4.6 Thinking ~1500, Gemini 3.1 Pro ~1493, GPT-5.4 ~1484.

Three things are analytically significant here. First, the frontier has gained 406 Elo points in three years — from 1,094 to 1,500. This is not incremental progress. It is a structural step-change with no parallel in previous enterprise software cycles. Second, Microsoft's presence at the capability frontier has been minimal: two months at #1 in late 2023, none since. Third: Copilot, deployed to millions of enterprise users as their primary AI tool, does not appear in Arena top rankings. Its value proposition was never capability competition. It was integration convenience.

None of this is a criticism of Microsoft's product strategy, which is internally coherent. It is a challenge to the enterprise strategies that read "we have Copilot" as equivalent to "we are AI-capable."

+406
Arena Elo points gained by leading models in 36 months — from 1,094 to 1,500
BenchLM.ai, May 2023 – April 2026
2/36
Months Microsoft held the #1 Arena position. OpenAI held it 16 months, Google 7.
BenchLM.ai Crown Change Tracker
8%
Copilot's active usage share when employees have access to multiple AI tools simultaneously
Recon Analytics, January 2026
68%
Copilot adoption when it is the only available tool — provisioning, not preference
Recon Analytics, January 2026

The Recon Analytics adoption data reinforces the capability story through a different lens. When employees have simultaneous access to Copilot and other tools, Copilot's active usage share falls to 8%. When it is the only available tool — the situation in most enterprise deployments — adoption reaches 68%. The 60-point gap between those numbers is a revealed preference. Given a choice, the large majority choose something else. This is the Siebel signal, early.

The Compound Nobody Is Measuring

The capability gap at first prompt is visible and increasingly difficult to ignore. The more consequential gap is the one that emerges through iteration — and it is almost entirely unmeasured in enterprise AI evaluations.

AI tools are not used once. Real knowledge work involves iteration: a prompt, a refinement, a restructure, a follow-up, a revision pass, a second restructure. Each iteration is an opportunity for a capable model to compound quality upward, or for a less capable model to plateau. The trajectories that look similar at iteration one look structurally different at iteration ten.

The Asymmetric Opportunity

The gap between what most enterprises deployed and what is actually possible creates an asymmetric opportunity — and this is where the next wave of economic value gets built.

Where the Gap Creates Opportunity

The Knowledge Asymmetry

Individual employees know which tool produces better output. Enterprises rarely measure this. The gap creates internal AI champions who will push procurement decisions — or leave for organisations that already made them.

The Iteration Compound

Teams using best-of-breed tools produce compoundingly better work with each successive iteration. Over 12 months this translates into a structural capability differential. Every workflow cycle widens the distance.

The Output Vacuum

Organisations that cut headcount on AI promises while deploying the wrong tool have created a gap between expected and delivered output. That vacuum gets filled — externally, flexibly, with better tools.

The Integration Convergence

The Copilot moat — native M365 integration — is a shrinking advantage. AI-first tools are adding integrations every quarter. The window where bundled AI's convenience advantage outweighs its capability deficit is closing.

The dot-com era, the mobile era, and every major platform shift before them followed a similar pattern: the asymmetry of access to new capability was large enough that the gap itself became a business model. The organisations that built in that gap — rather than waiting for enterprise procurement cycles to resolve it — were the ones that mattered at the end of the decade.

The gap between what was deployed and what's actually possible is where dot-com v2 gets built. Not inside the enterprise. Outside it first. Then back in — as the capability that incumbents must acquire or replicate under competitive pressure.

What This Means in Practice

To be precise: enterprise constraints are real. Compliance requirements, data governance, security architecture, vendor relationships — these are legitimate structural realities, not bureaucratic friction. Copilot solves genuine problems within the M365 ecosystem and its integration advantages are not trivial.

The argument is not "abandon Copilot." The argument is that enterprise AI strategy needs to distinguish between two questions that are currently being conflated:

Question 1: What AI tool is safe, compliant, and deployable at scale within our existing infrastructure? Copilot is often a reasonable answer.

Question 2: What AI capability do we need to be competitively positioned in 2027? Copilot is rarely a sufficient answer — and the gap between those two answers is where the strategic risk lives.

Treating the answer to Question 1 as a sufficient answer to Question 2 is the Siebel error. It is being made at scale, in real time, across most large enterprises. And like the Siebel error, it compounds quietly — iteration by iteration, sprint cycle by sprint cycle — until the gap is too large to close incrementally.

Recommendations

  1. Audit your capability ceiling, not just your adoption rate. Adoption metrics measure distribution. Capability metrics measure what's actually possible. Run controlled output comparisons on your real use cases — not vendor demos. The most revealing test: same prompt, three tools, ten iterations. Most enterprises have never done this.

  2. Decouple compliance infrastructure from capability strategy. The answer to "what can we deploy securely" and "what should our people be capable of" do not need to be the same tool. Hybrid models — best-of-breed for high-value capability workflows, bundled for integrated productivity — are operationally viable and strategically superior to a single-tool default.

  3. Measure the iteration compound, not the first output. When evaluating AI tools for strategic workflows, the relevant metric is not output quality at prompt one. It is output quality at prompt ten, after a real workflow with real refinement. The compound gap is the strategically significant one — and it is currently absent from most enterprise AI evaluations.

  4. Watch revealed preference, not stated preference. The tools your best people choose when nobody is watching are a leading indicator of where capability is moving. When given access to multiple tools, 92% of employees do not actively choose Copilot. This is a signal worth acting on before it becomes a retention and competitive capability problem simultaneously.

The Bigger Picture

The bundled vs best-of-breed debate in AI is not a niche technology procurement question. It is a strategic positioning question that will differentiate organisations over the next three years in ways that are currently underestimated.

The organisations that close this gap — not by abandoning their infrastructure constraints, but by refusing to let those constraints define their capability ceiling — will look structurally different in 2028 from the ones that didn't. The ones that wait for their bundled vendor to close the capability gap are making the same bet Siebel's customers made in 2003.

Some of them were right. Most were not.


Sources: Chatbot Arena / LMSYS · BenchLM.ai Arena Elo Tracker · Recon Analytics — AI Choice 2026 · Oracle layoffs — CNBC · WizardLM — Microsoft Research