Cookie Consent

I use cookies to understand how my website is used. This data is collected and processed directly by me, not shared with any third parties, and helps us improve our services. See my privacy and cookie policies for more details.

The Inference Migration: What Consumer Agents Mean for Enterprise AI's Next Phase

New York | Published in AI and Board | 12 minute read |    
A corporate boardroom table overrun with small, friendly red robotic lobsters with glowing blue eyes, perched on laptops, documents, and coffee cups, with a city skyline visible through floor-to-ceiling windows and business charts displayed on a presentation screen (Image generated by ChatGPT 5.2)

The demand signal for AI’s next infrastructure phase is hiding in plain sight — not in enterprise contracts or analyst projections, but in what consumers are voluntarily paying for always-on AI assistance. OpenClaw, an open-source AI agent that evolved through Clawdbot, then Moltbot, before settling on its current name in January, has accumulated more than 100,000 GitHub stars within weeks of launch, making it one of the fastest-growing open-source projects in recent memory. What matters for Boards isn’t the star count — it’s the economics. Heavy users report spending $10–25 per day on API costs to keep their personal agent running continuously, translating to $3,650–9,125 annually for what users describe as a chief of staff that never sleeps.

The combined annual cost of a Netflix subscription, Spotify premium, a TV package, and a mobile phone contract sits comfortably under $2,000. Early adopters of always-on AI agents are spending three to six times that on inference alone. Consumers are paying more for AI inference than for their entire entertainment stack. When ChatGPT launched in November 2022, it became the fastest-growing application in history — and within eighteen months, shadow enterprise adoption had surged, outpacing most organisations’ governance frameworks. OpenClaw represents the next iteration of that pattern. The infrastructure case for a world where professionals maintain persistent AI agents isn’t speculative. Consumers are already paying for it.

The architectural shift: from diurnal to always-on

ChatGPT and its contemporaries operate in episodic, query-response patterns — a human initiates a prompt, the model responds, and the cycle ends, with the system sitting idle between exchanges. This mirrors how most enterprise AI deployments work today: discrete tasks, bounded sessions, clear start and end points.

OpenClaw and its emerging cohort represent something architecturally different. These are persistent agents that run continuously, maintaining context across days and weeks of interaction. They execute real tasks — web browsing, email management, calendar coordination, file operations, code assistance — not merely answering questions but completing workflows autonomously. Connection via WhatsApp, Telegram, Signal, and iMessage means the agent is always reachable, always accumulating context, always ready to act.

The inference economics change fundamentally with this architecture. Traditional AI usage follows human diurnal patterns — queries cluster during working hours and drop to near zero overnight. Always-on agents invert this entirely, running continuous inference while the user sleeps, works, and commutes. This isn’t eight to ten hours of active use but twenty-four hours of persistent availability, multiplying inference demand by a factor that current infrastructure wasn’t designed to accommodate.

The market arithmetic is significant. Ten million users at $20 per day generates approximately $73 billion annually in consumer inference spend; scale to one hundred million users — a modest penetration for a genuinely compelling utility — and the figure approaches $730 billion annually at full adoption. For context, the global streaming market is worth approximately $100 billion and global mobile subscriptions approximately $500 billion.

Amazon CEO Andy Jassy’s observation last week that “as fast as we install this AI capacity, it’s getting monetized” becomes comprehensible against this arithmetic. This isn’t speculation about what consumers might eventually pay — it’s observation of what early adopters already pay.

The ChatGPT precedent: a pattern that predicts

The consumer-to-enterprise adoption pipeline has a well-documented precedent. ChatGPT launched as a consumer novelty in November 2022. By 2023, shadow enterprise adoption had surged — Menlo Security documented a 68% increase in ungoverned AI usage within enterprises, with 90% of employees using AI tools outside formal controls. By 2024, formalisation accelerated — Deloitte reported 60% of workers now had access to sanctioned AI tools. By 2025–26, scaled deployment with nascent governance frameworks had become standard practice for leading organisations.

The pattern is consistent: consumer adoption served as leading indicator for enterprise demand, with shadow usage — employees bringing consumer tools into work — bridging the gap between consumer enthusiasm and enterprise readiness. Formalisation followed once organisations recognised that governing usage was preferable to ignoring it, and the entire cycle from consumer novelty to enterprise infrastructure took approximately three years.

Applying this pattern to agentic AI produces a timeline Boards should take seriously. 2025–26 represents consumer agentic adoption — the OpenClaw phenomenon is this phase made visible. 2026 will likely see shadow agentic AI arriving in enterprises as employees connect personal agents to work email, work calendars, and work documents. 2027 should see formalised enterprise agentic platforms from hyperscalers and enterprise software vendors reach broad adoption. 2028 and beyond will see scaled agentic deployment with mature governance frameworks.

Infrastructure providers watched the ChatGPT cycle unfold — consumer demand predicting enterprise demand with remarkable fidelity. Current capital expenditure commitments, with 2026 analyst projections exceeding $600 billion combined for major hyperscalers and approximately 75% directed at AI, reflect pattern-matching against a known demand curve. Organisations that dismissed ChatGPT as consumer novelty in late 2022 found themselves scrambling to govern shadow AI by mid-2024. The same window is now opening for agentic systems.

The determinism question: why the objection is narrower than it appears

Boards encountering agentic AI for the first time frequently raise a legitimate concern: consumer AI tolerates errors, but enterprise processes require determinism and reliability. This objection deserves serious examination — and closer inspection reveals it addresses a narrower slice of enterprise work than commonly assumed.

Much of knowledge work involves what might be called soft processes: drafting, research, synthesis, communications, analysis, and creative development. These processes tolerate — and frequently benefit from — nondeterministic AI outputs. A draft email refined through AI assistance doesn’t need to be identical every time; it needs to be effective. The verification premium is what matters here: skilled professionals guide the AI and validate its outputs, producing better results than either could alone. BCG’s research confirms this, showing that 67% of employees in companies that redesigned workflows around AI save over an hour daily, compared with 49% in organisations that merely deployed tools without rethinking how work happens.

Hard processes — regulatory compliance, financial reporting, audit trails, safety-critical operations — do require precision and determinism, and they are often high-stakes. Yet even here, the question isn’t whether to exclude AI but how to architect human-AI collaboration appropriately.

The enterprise opportunity lies not in replacing deterministic systems but in wrapping them with predictive intelligence. Consider an ERP system that executes transactions versus the same ERP wrapped with demand forecasting, scenario modelling, and real-time optimisation. The underlying deterministic processes don’t break — they become competitively inadequate when rivals wrap them with intelligent orchestration. The compound-loop model makes clear why: agents can invoke any AI discipline — generative models, classical machine learning, predictive analytics, rules engines, robotic process automation, and existing business processes. Agents don’t replace deterministic systems — they coordinate them within intelligent workflows.

This creates a three-stage ramp. Consumer inference operates with high tolerance for nondeterminism — personal productivity, creative assistance, research. Enterprise soft processes operate with moderate tolerance — drafting, synthesis, communications, and analysis that constitute the majority of knowledge work. Enterprise hard processes operate with low tolerance — compliance, financial reporting, and safety-critical operations that benefit from predictive layers wrapping deterministic execution.

The determinism objection, when examined carefully, addresses a narrower slice of enterprise use cases than commonly assumed. The remainder follows the same adoption curve as consumer applications.

Shadow agentic AI: the governance imperative

If the ChatGPT precedent holds — and every indicator suggests it will — shadow agentic AI is already arriving in enterprises. Employees with personal OpenClaw instances are connecting them to work email, calendars, and documents. API keys for corporate accounts are flowing through personal agent configurations. Security researchers have documented vulnerabilities including exposed API keys and prompt injection risks — attack surfaces most enterprise security frameworks weren’t designed to detect.

Most enterprise AI policies still address only episodic chatbots. They were never designed for agents that maintain context across months of continuous operation. Existing controls assume a human initiates each interaction, not that an autonomous agent monitors and acts on enterprise data streams around the clock. Data governance frameworks don’t anticipate AI systems accumulating institutional knowledge that may exceed what any individual employee possesses.

The shadow AI dynamics documented by Menlo Security — 90% using AI outside enterprise controls — and BCG — 54% willing to use unauthorised tools when corporate solutions fall short — will inevitably manifest in agentic form. When 52% of AI users already save more than an hour daily, the productivity incentive to connect a personal agent to work systems is overwhelming.

The timeline for action is compressed. By H2 2026, shadow agentic AI will be spreading through organisations. By 2027, formalised enterprise agentic platforms will mature — and organisations without governance frameworks will scramble to retrofit controls onto adoption already underway. By 2028, organisations that established minimum lovable governance early will capture compound advantage over those still building foundations.

Minimum lovable governance for agentic AI means proportionate controls matching agent capability to risk profile, clear policies on persistent agent access, monitoring for unauthorised API connections from personal tools, and pathways to convert shadow agentic usage into governed capability — the amnesty model applied to agents. Critically, this governance must be embedded in workflows rather than imposed as a separate compliance burden. The organisations that acted early on generative AI governance now have frameworks that enable rather than constrain innovation. The same window is opening for agentic governance.

The energy multiplier: infrastructure implications

The shift from episodic to always-on inference carries infrastructure implications that extend well beyond compute capacity. The diurnal patterns that allow current infrastructure to breathe — active hours with overnight lulls — disappear entirely when agents run continuously, forcing provisioning for sustained, flat-line demand that fundamentally changes the economics of AI capacity planning.

For UK Boards, the energy dimension compounds the challenge. UK electricity costs run approximately 4× US rates for industrial consumers, which means a UK knowledge worker’s always-on agent costs roughly four times what an equivalent American worker’s agent costs — compounding across every employee, every day. As AI agents transition from novelty to productivity utility, this differential creates a structural competitiveness gap. The organisations controlling energy infrastructure for AI compute — hyperscalers moving toward micro-generation and long-term power purchase agreements — become strategic chokepoints.

Energy policy becomes AI policy when always-on inference scales. Grid infrastructure determines who can afford persistent intelligence, and the always-on shift multiplies energy demands at precisely the moment that cost differentials create competitive asymmetries. For UK Boards, the question now extends beyond AI governance to energy strategy.

Strategic implications for Boards

Seven considerations deserve Board attention. Consumer adoption is now a leading indicator — tracking consumer AI spending reveals enterprise demand trajectory more reliably than analyst projections. The market arithmetic reinforces this: $730 billion in potential consumer inference spend at full adoption validates infrastructure investment theses and signals the scale of what follows for enterprise.

The determinism objection, as examined above, is narrower than it appears — most knowledge work tolerates nondeterminism, and hard processes benefit from intelligent wrapping rather than replacement.

Shadow AI signals demand, not defiance. When 54% of employees say they will use unauthorised tools and 90% already operate outside controls, the question isn’t whether AI spreads but how it’s governed. AI wrappers compound the pressure — existing processes don’t fail, they become inadequate when competitors wrap them with predictive intelligence. Governance frameworks are needed now, because the 2026–2027 window determines whether organisations capture agentic value through strategy or discover it through crisis. And energy policy becomes AI policy — always-on inference makes energy costs strategic differentiators that Boards can no longer treat as operational background noise.

The strategic sequence runs from consumer agents — happening now — through enterprise generative AI, which is largely complete, to agentic orchestration of deterministic systems in 2027–28. Each phase builds on the previous; organisations that missed generative adoption face steeper climbs to agentic capability. McKinsey’s 2025 research underscores the gap: 88% of organisations now use AI in at least one function, but only 23% have successfully scaled agentic systems. PwC’s Global AI Jobs Barometer quantifies the stakes — AI-exposed industries show 3× higher revenue per employee growth and a 56% wage premium for AI skills.

The choice is becoming familiar: establish minimum lovable governance for agentic AI now, positioning for the 2027 formalisation wave, or discover shadow agentic AI spreading by H2 2026 and scramble to retrofit governance on adoption already underway.

Conclusion

Consumer willingness to pay $3,650–9,125 annually for AI inference — more than their combined entertainment subscriptions — represents revealed preference that enterprise demand will follow. The ChatGPT precedent demonstrated this pipeline: consumer novelty became shadow enterprise adoption became formalised deployment within three years. OpenClaw signals the same trajectory for agentic systems, with a critical difference — the shift from episodic to always-on inference changes capacity planning, energy requirements, and governance assumptions simultaneously.

The real opportunity is compound architecture. Agents that orchestrate deterministic systems don’t replace existing processes — they wrap them with predictive intelligence that creates competitive advantage. Minimum lovable governance established now positions organisations to capture agentic value when formalised platforms arrive in 2027. Governance retrofitted after shadow adoption constrains rather than enables.

The ChatGPT governance cycle took three years from consumer adoption to enterprise infrastructure. The agentic cycle is running on a similar timeline, and 2026 is the year for governance frameworks. By 2027, the consumer-to-enterprise pipeline will deliver shadow agentic AI whether organisations are prepared or not.

The inference migration is underway. 2026 is the narrow window to establish minimum lovable governance before the shadow agentic wave arrives. The question is whether your governance keeps pace.

Let's Continue the Conversation

Thank you for reading about consumer AI adoption patterns and what they signal for enterprise demand. I'd welcome hearing about your Board's perspective on agentic AI governance — whether you're already seeing employees adopt personal AI agents, developing governance frameworks in anticipation of the shadow agentic wave, or wrestling with how to position for the 2027 formalisation phase while maintaining competitive agility.