Cookie Consent

I use cookies to understand how my website is used. This data is collected and processed directly by me, not shared with any third parties, and helps us improve our services. See my privacy and cookie policies for more details.

Tagged with: #decision-making

Posts tagged with #decision-making explore how to reshape your approach to strategic and operational decisions through responsible AI adoption.

Maximum Fidelity: How Four Indicator Types Strengthen Board Decisions

New York | Published in Board | 13 minute read |    
A close-up photograph of a professional audio mastering console, showing a warmly lit analogue VU meter on the left with its amber-glowing face, flanked by precision control knobs and monitoring switches on a dark panel. The shallow depth of field draws the eye to the meter itself, with the surrounding controls falling gently into shadow. An image representing the precision instruments used by audio engineers to measure fidelity, used here as a metaphor for the four indicator types that give Boards maximum fidelity on the decisions in front of them (Image generated by ChatGPT 5.4)

Boards have always governed under incomplete information. What the four indicator types offer is not more information but a progressively higher quality of it. Lagging indicators establish what happened, leading indicators signal direction, predictive indicators model possible futures, and reasoned indicators prove what is certain. Applied in combination to a single decision, they represent maximum fidelity — everything knowable and made available before the judgement is made. This article explains why the distinction between a decision made with maximum fidelity and one made without it matters for every director around the table.


From Probable to Provable: What Automated Reasoning Means for the Board

Washington DC | Published in Emerging | 13 minute read |    
A geometric wire-frame lattice structure resting on architectural blueprints, surrounded by drafting tools, symbolising the formal constraints and mathematical rigour that underpin automated reasoning (Image generated by ChatGPT 5.4)

Boards have always governed under conditions of incomplete information. What has changed is the volume and velocity of that information, and the speed at which AI systems now act upon it. Lagging indicators report on the past. Leading indicators signal what is likely to happen next. Predictive indicators model possible futures. But automated reasoning offers something different entirely: proof. Not a tighter estimate, but a formally verified property of the decision space itself. This article explains what automated reasoning is, where it already operates across regulated industries, and why it represents a new class of governance instrument for Boards.


AI and the Director: A Practical Playbook for Governing What You Can't Fully See

London | Published in Board | 11 minute read |    
A figure in a dark suit, partially concealed behind a heavy charcoal velvet curtain, one hand gripping the curtain edge in sharp directional light against a black background — a visual metaphor for the unseen operator whose workings a director is expected to trust without seeing. (Image generated by ChatGPT 5.4)

The informational asymmetry between management and the Board has always been the central tension of governance. For AI, it is no longer manageable through existing structural checks; the distance is not merely larger than previous technology waves, it is qualitatively different. A director must be able to interrogate maturity claims, assess whether governance is operational or merely presentational, and identify which AI risks are personal development challenges and which are failures of oversight itself. The IoD has formally named the gap. This article defines what closing it actually requires: not technical fluency, but specific capacities for independent evaluation mapped against the governance obligations every director carries, and a diagnostic framework for identifying exactly where the work needs to start.


The Great Remaking: The Questions Boards Should Be Asking About Their AI Position

Llantwit Major | Published in AI | 10 minute read |    
Aerial view of a landscape as clouds gradually clear, with sunlight revealing the underlying terrain, representing how a board-level diagnostic cuts through activity metrics to expose the organisation’s true AI position (Image generated by ChatGPT 5.2)

The part of AI value that is technological and replicable is also the part that standard progress measures capture best. Pilot counts, budget lines, and strategy documents say nothing about whether the essence of work is genuinely being remade, or whether the three compounding loops are operating. A Board that accepts those reports without probing them is not exercising oversight; it is ratifying a narrative the evidence shows is inflated. This article provides the diagnostic that does: probing questions structured around the data, talent, and process redesign loops, with an interpretive guide to what credible answers look like — and what their absence reveals.


MCP Explained: The Agent Infrastructure Standard Boards Need to Understand

Llantwit Major | Published in Data | 11 minute read |    
A sleek modern MCP hub on a dark walnut executive desk, with cables of different vintages connecting to surrounding legacy hardware including a CRT monitor, blue LED glowing on the hub. (Image generated by ChatGPT 5.2)

An AI agent that can only see the public internet is no more useful to an organisation’s business than a very expensive search engine. The intelligence is not the constraint. The connectivity is. Model Context Protocol — MCP — is the infrastructure standard that connects agents to the proprietary data, systems, and processes that constitute real competitive advantage. This article explains what MCP is, why the major enterprise vendors have already converged on it, and the governance questions Boards should be asking before their technology teams answer them by default.


The Verification Premium: What Classical Training Reveals About AI Coding Costs

New York | Published in AI and Board | 13 minute read |    
My desktop setup: reMarkable Paper Pro for ideas, MacBook Air M2, a headless NVIDIA DGX Spark handling the heavy lifting, and the tools behind the experiment — Amazon Kiro, Claude Code, and a terminal window. Plus the late-night lighting that makes it feel like coding in the 1980s again.

AI coding tools don’t close the expertise gap — they amplify it. Research shows senior developers capture twice the productivity gains of juniors, while a randomised controlled trial found experienced developers actually worked slower with AI than without, the hidden taxes of verification offsetting initial speed. This article explores the verification premium — and why Boards should ask not “can we use AI to write code cheaper?” but “do we have the verification capability to ensure AI-generated code creates value rather than debt?”


World Models: The Next Horizon in AI for Predictive Enterprise Intelligence

Llantwit Major | Published in AI , Board and Emerging | 10 minute read |    
Executives in a strategic command centre viewing holographic simulations of global operations with aircraft above a digital world map, representing predictive AI transforming reactive management into anticipatory strategy (Image generated by ChatGPT 5)

World models mark AI’s shift toward true predictive power, allowing systems to simulate future scenarios and help businesses move from reacting to events to anticipating them. Drawing on emerging research, including Yann LeCun’s work on simulation-based intelligence, this article highlights the practical gains industries like aviation and finance are seeing in operational efficiency through these future-looking tools. For Boards, world models present a tantalising future: the opportunity to turn future insight into present advantage.


The Accountability Gap: When AI Delegation Meets Human Responsibility

New York | Published in AI and Board | 15 minute read |    
Senior executives observing a fast-moving automated conveyor belt of AI-generated business reports in a modern corporate office, with unused quality control tools in the foreground illustrating the AI accountability gap (Image generated by ChatGPT 5)

While organisations transfer decision-making agency to AI systems, accountability remains with humans, yet boards approve AI deployment without investing in the verification capability needed to ensure it. In this article, I demonstrate why this creates a strategic choice with measurable consequences: augmentation preserves expertise pipelines whilst achieving efficiency gains, but replacement destroys capabilities that cannot be rebuilt, turning apparent cost reduction into systematic competitive disadvantage.


Agentic AI: Strip Away the Hype and Understand the Real Strategic Choice

Llantwit Major | Published in AI and Board | 17 minute read |    
Modern corporate boardroom scene split between thoughtful business executives on the left working with documents representing human-in-the-loop decision-making, and multiple glowing AI agent representations on the right operating autonomously in parallel, symbolising the strategic choice about where to transfer agency from humans to machines (Image generated by ChatGPT 5)

Agentic AI has become this year’s poster child, dethroning generative AI as the technology everyone wants to discuss. Yet fundamental misunderstandings about what agentic systems actually do create barriers to successful adoption. This article demystifies the hype by revealing the core truth: agentic AI is generative AI in a loop, where the machine drives iteration instead of a human, making the strategic question not about technology sophistication but where to consciously transfer decision-making agency from people to systems, and at what scale.


From AI Pilots and Projects to AI Strategy: Avoiding the Business Case Trap

Sydney | Published in AI and Board | 10 minute read |    
Multiple small groups of musicians scattered across a grand concert hall, each playing different pieces of music simultaneously, creating fragmentation despite individual excellence (Image generated by ChatGPT 5)

Boards are approving AI initiatives at record pace – 92% of companies plan increased investment – yet only 1% have achieved AI maturity: the gap reveals a fundamental misconception about AI strategy. In this article, I expose why accumulating business cases creates fragmentation rather than transformation, and why Boards must shift from project-level approvals to orchestrating systematic AI capability before their disconnected pilots become an expensive collection of failures.