<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Marco's Substack]]></title><description><![CDATA[Enterprise AI adoption, Agentic systems, and the operations that make them work.]]></description><link>https://marcosnewsletter.substack.com</link><image><url>https://substackcdn.com/image/fetch/$s_!eppd!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F4f4d75f5-0eca-4493-a3d3-c318edf2ee4c_648x648.png</url><title>Marco&apos;s Substack</title><link>https://marcosnewsletter.substack.com</link></image><generator>Substack</generator><lastBuildDate>Sun, 17 May 2026 03:09:12 GMT</lastBuildDate><atom:link href="https://marcosnewsletter.substack.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Marco]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[marcosnewsletter@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[marcosnewsletter@substack.com]]></itunes:email><itunes:name><![CDATA[Marco]]></itunes:name></itunes:owner><itunes:author><![CDATA[Marco]]></itunes:author><googleplay:owner><![CDATA[marcosnewsletter@substack.com]]></googleplay:owner><googleplay:email><![CDATA[marcosnewsletter@substack.com]]></googleplay:email><googleplay:author><![CDATA[Marco]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Person-Led AI Has a Ceiling. Operational AI Has Compounding.]]></title><description><![CDATA[Last week I split AI agents into Person-Led AI and Operational AI. This week, the deeper question: who runs the task. The difference decides whether your AI investment compounds or evaporates.]]></description><link>https://marcosnewsletter.substack.com/p/person-led-ai-ceiling-operational-ai-compounding</link><guid isPermaLink="false">https://marcosnewsletter.substack.com/p/person-led-ai-ceiling-operational-ai-compounding</guid><dc:creator><![CDATA[Marco]]></dc:creator><pubDate>Wed, 13 May 2026 22:35:40 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!FQHU!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F930f16d1-ecf0-4842-aa8b-dce78e8a9566_1730x909.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>The conversation could have ended there &#8594; Most professional conversations about AI in 2026 stop at &#8220;we use it for X and we are seeing gains&#8221;.</p><p>The gain is real. The conversation is incomplete.</p><p>What I tried to add to that exchange is the distinction that decides whether AI compounds inside an organization or stays trapped inside the people who use it. Most enterprise leaders have not yet been given clean language for it.</p><p>That distinction is not about the tool. It is about who runs the task.</p><p>When you open the computer to ask Claude to draft a specification, you have run the task. The AI helped, you guided, the output is yours. This is Person-Led AI. It is real, it is valuable, and it has a ceiling.</p><p>When the agent runs the task on its own, on a configured trigger, and contacts you only when it needs your specific input, you have stopped running the task. This is Operational AI. It is harder to set up. It compounds.</p><blockquote><p>Person-Led AI is real, valuable, and has a ceiling. Operational AI is harder to set up, and it compounds.</p></blockquote><p>Both are legitimate. Both are useful. They are not interchangeable, and the difference will explain a large share of the AI value capture gap that enterprise reports keep documenting.</p><div><hr></div><h2>The operational definition of who runs the task</h2><p>The simplest test for any AI deployment is the question of who initiates.</p><p>In Person-Led AI, a person opens the computer, writes a prompt, attaches a file, asks for help. The trigger is human and ad hoc. The decision to engage the AI is made fresh each time, by the person who needs the work done. The work happens because someone chose, in that moment, to engage.</p><p>In Operational AI, the agent starts on its own. The trigger is configured. It can be temporal (every Monday morning), event-driven (a new ticket arrives), threshold-driven (a metric crosses a value), or systemic (another system signals a state change).</p><p>No one decides, in that moment, to engage the AI. The AI was already engaged.</p><p>This sounds like a small distinction. It is not. The shift from &#8220;a person initiates&#8221; to &#8220;a trigger initiates&#8221; changes the economics, the operating model, the failure mode, and the kind of value that gets created.</p><p>Take the monthly business review as a concrete example.</p><p>In a Person-Led AI setup, on the first of the month a finance analyst opens her tools, queries the data warehouse, pulls metrics into a deck, drafts narratives, sends to her manager for review. Claude or Copilot speeds up each of these steps. The analyst is faster. The work still depends entirely on her starting it.</p><p>In an Operational AI setup, on the first of the month an agent runs. It pulls the same data, generates the same draft, and contacts the analyst with specific questions where it needs domain expertise: a variance it cannot explain, a one-off transaction that breaks pattern, a narrative judgment call. The analyst answers the specific questions. The agent finishes the deck and routes it.</p><blockquote><p>Same output. Same monthly cadence. Two profoundly different operating models. When the analyst leaves, the first model leaves with her. The second model survives.</p></blockquote><p>In the first, the value of the work scales with the analyst. In the second, the value scales with the system.</p><div><hr></div><h2>Person-Led AI is real, useful, and bounded</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!FQHU!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F930f16d1-ecf0-4842-aa8b-dce78e8a9566_1730x909.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!FQHU!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F930f16d1-ecf0-4842-aa8b-dce78e8a9566_1730x909.png 424w, https://substackcdn.com/image/fetch/$s_!FQHU!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F930f16d1-ecf0-4842-aa8b-dce78e8a9566_1730x909.png 848w, https://substackcdn.com/image/fetch/$s_!FQHU!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F930f16d1-ecf0-4842-aa8b-dce78e8a9566_1730x909.png 1272w, https://substackcdn.com/image/fetch/$s_!FQHU!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F930f16d1-ecf0-4842-aa8b-dce78e8a9566_1730x909.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!FQHU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F930f16d1-ecf0-4842-aa8b-dce78e8a9566_1730x909.png" width="1456" height="765" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/930f16d1-ecf0-4842-aa8b-dce78e8a9566_1730x909.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:765,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2609683,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://marcosnewsletter.substack.com/i/196538566?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F930f16d1-ecf0-4842-aa8b-dce78e8a9566_1730x909.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!FQHU!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F930f16d1-ecf0-4842-aa8b-dce78e8a9566_1730x909.png 424w, https://substackcdn.com/image/fetch/$s_!FQHU!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F930f16d1-ecf0-4842-aa8b-dce78e8a9566_1730x909.png 848w, https://substackcdn.com/image/fetch/$s_!FQHU!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F930f16d1-ecf0-4842-aa8b-dce78e8a9566_1730x909.png 1272w, https://substackcdn.com/image/fetch/$s_!FQHU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F930f16d1-ecf0-4842-aa8b-dce78e8a9566_1730x909.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"></div></div></a></figure></div><p>I want to be careful here, because the easy mistake is to read this as &#8220;Person-Led AI is bad, Operational AI is good&#8221;. That is not what I am saying.</p><p>Person-Led AI is generating real, measurable value at enterprise scale. The numbers are unambiguous.</p><p>According to <a href="https://menlovc.com/perspective/2025-the-state-of-generative-ai-in-the-enterprise/">Menlo Ventures&#8217; 2025 State of Generative AI</a>, copilots dominate the horizontal AI market: $7.2 billion of an $8.4 billion total, or 86%. Microsoft Copilot, Claude for Work, ChatGPT Enterprise, and adjacent tools are responsible for the majority of enterprise AI spend in 2025-2026.</p><p>They are responsible because they work.</p><p>A developer using Copilot Workspace ships features faster. A finance analyst using Claude drafts variance commentary in minutes instead of hours. A marketing manager using ChatGPT Enterprise iterates on positioning across audiences before lunch.</p><p>These are real productivity gains, captured by individuals, that compound at the level of individual output. According to <a href="https://learnexperts.ai/blog/linkedin-workplace-learning-report/">LinkedIn&#8217;s Workplace Learning Report</a>, AI literacy is the number one rising skill on the platform in 2025.</p><p>Person-Led AI is not a fad. It is the new baseline.</p><p>But it has a ceiling.</p><p>Person-Led AI scales linearly with the people who use it. If you have 500 employees who each save 30 minutes a day with Copilot, you have created 250 hours of daily productivity. Real, measurable, valuable.</p><p>Now multiply: how do you get to 2,500 hours? You hire ten times more people. The model is per-seat, the value is per-person, the ceiling is your headcount.</p><p>McKinsey&#8217;s <a href="https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai">State of AI 2025</a> puts a sharper edge on this. Only 6% of organizations qualify as high performers, defined as those generating 5% or more EBIT impact from AI.</p><p>The 94% that do not capture meaningful value are not, in most cases, organizations that failed at Person-Led AI. They are organizations that succeeded at Person-Led AI and stopped there. They bought copilot licenses, their people got faster, and the financial metrics did not move.</p><p>The gap between individual productivity and enterprise value is the gap between Person-Led AI and what comes next.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://marcosnewsletter.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">If this is the kind of distinction you want to keep working with, consider subscribing. I publish data-driven breakdowns on enterprise AI maturity every Tuesday, no hype, no vendor pitches.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><h2>Operational AI is where the work runs without you</h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!YCQ-!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b58c57d-b121-450a-b957-2fd63fce36b0_1730x909.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!YCQ-!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b58c57d-b121-450a-b957-2fd63fce36b0_1730x909.png 424w, https://substackcdn.com/image/fetch/$s_!YCQ-!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b58c57d-b121-450a-b957-2fd63fce36b0_1730x909.png 848w, https://substackcdn.com/image/fetch/$s_!YCQ-!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b58c57d-b121-450a-b957-2fd63fce36b0_1730x909.png 1272w, https://substackcdn.com/image/fetch/$s_!YCQ-!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b58c57d-b121-450a-b957-2fd63fce36b0_1730x909.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!YCQ-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b58c57d-b121-450a-b957-2fd63fce36b0_1730x909.png" width="1456" height="765" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3b58c57d-b121-450a-b957-2fd63fce36b0_1730x909.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:765,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2766594,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://marcosnewsletter.substack.com/i/196538566?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b58c57d-b121-450a-b957-2fd63fce36b0_1730x909.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!YCQ-!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b58c57d-b121-450a-b957-2fd63fce36b0_1730x909.png 424w, https://substackcdn.com/image/fetch/$s_!YCQ-!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b58c57d-b121-450a-b957-2fd63fce36b0_1730x909.png 848w, https://substackcdn.com/image/fetch/$s_!YCQ-!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b58c57d-b121-450a-b957-2fd63fce36b0_1730x909.png 1272w, https://substackcdn.com/image/fetch/$s_!YCQ-!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F3b58c57d-b121-450a-b957-2fd63fce36b0_1730x909.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"></div></div></a></figure></div><p>Operational AI is harder to build, harder to govern, and harder to explain to a budget committee. It is also where the value compounds.</p><p>The pattern is not theoretical. It is in production now, at scale, in companies most enterprise leaders have heard of.</p><p>Salesforce Agentforce, in nearly all its production deployments, is event-triggered. An incoming email, a form submission, a sensor alert kicks off a multi-step workflow that the agent runs to completion or to escalation. According to <a href="https://www.innovadeltech.com/salesforce-agentforce-implementation-guide/">implementation data from the field</a>, well-configured deployments achieve 40-60% deflection on routine inquiry categories.</p><p>The customer never knows a human did not handle it. No human did.</p><p>GitHub Copilot Coding Agent, <a href="https://github.blog/news-insights/product-news/github-copilot-meet-the-new-coding-agent/">generally available since 2025</a>, inverts the developer relationship with AI. You assign an issue to the agent. The agent works in the background, writes code, runs tests, opens a pull request. You review later.</p><p>Early adopters report 30-70% productivity gains for routine development tasks. The trigger is the issue assignment. The work runs autonomously.</p><p>Datadog&#8217;s <a href="https://www.itprotoday.com/it-operations/datadog-dash-2025-autonomous-agents-transform-it-observability-operations">Bits AI suite</a> deployed an autonomous SRE Agent that performs hypothesis-driven root cause analysis in under a minute. It runs continuously, triggered by anomalies in observability data. Datadog claims thousands of internal engineering hours saved per week.</p><p>No on-call engineer initiates the agent&#8217;s work. The infrastructure does.</p><p>Financial services may be the most aggressive adopter. According to <a href="https://aws.amazon.com/blogs/awsmarketplace/agentic-ai-solutions-in-financial-services/">AWS data on agentic AI in financial services</a>, 53% of financial institutions are running AI agents in production in 2025. Trades execute on signal, not on a trader&#8217;s command. Risk models adjust on threshold breaches, not on a quant&#8217;s morning coffee.</p><p>The reported ROI is $3.50 per $1 invested, with 35% cost reduction and 55% higher operational efficiency.</p><p>The shift is happening fast. Gartner predicts that <a href="https://www.gartner.com/en/newsroom/press-releases/2025-08-26-gartner-predicts-40-percent-of-enterprise-apps-will-feature-task-specific-ai-agents-by-2026-up-from-less-than-5-percent-in-2025">40% of enterprise applications will feature embedded AI agents by end of 2026</a>, up from less than 5% in 2025.</p><p>Menlo Ventures reports that in Q1 FY2026 alone, 160,000 organizations created over 400,000 custom agents using Microsoft Copilot Studio. The agent platform market, currently around $750 million, is forecast to reach $80-100 billion by 2030 according to <a href="https://www.gartner.com/en/articles/hype-cycle-for-agentic-ai">Gartner&#8217;s Hype Cycle for Agentic AI</a>.</p><p>The ratio between Person-Led AI and Operational AI in enterprise spend will not hold.</p><p>The reason it will not hold is that Operational AI compounds in a way Person-Led AI cannot.</p><p>When the SRE Agent at Datadog catches a regression at 3 AM and resolves it before anyone wakes up, the value is captured by the system. When the Salesforce Agentforce instance handles 60% of routine tickets without human involvement, the value is captured by the workflow. When the financial agent rebalances on signal, the value is captured by the architecture.</p><p>None of this depends on a specific person being good at their job. The value sits in the infrastructure.</p><div><hr></div><h2>This is not a new framework, it is a translation</h2><p>Before I go further, a credit owed to people who have been thinking about this for decades.</p><p>The distinction between Person-Led AI and Operational AI maps onto a much older concept from automation engineering. <a href="https://ieeexplore.ieee.org/document/844354">Parasuraman and Sheridan</a>, in a foundational 2000 paper, distinguished between human-in-the-loop systems, where the human is a constant active participant, and human-on-the-loop systems, where the human supervises and intervenes only on escalation.</p><p>Their model has been the canonical reference for autonomy governance in safety-critical systems for 25 years.</p><p>What I am calling Person-Led AI is a strict version of human-in-the-loop, where the human is not just validating but executing every step. What I am calling Operational AI is human-on-the-loop with a particular twist: the agent contacts the human proactively, rather than the human checking in.</p><p>I am not proposing this as a discovery. I am translating a concept that automation engineers have used for a generation into language that lands in enterprise AI strategy meetings.</p><blockquote><p>Loop control authority does not survive a budget review. Who runs the task does.</p></blockquote><p>The distinction is the same. The vocabulary is calibrated for the room where the spend is approved.</p><p>I mention this for two reasons. First, intellectual honesty: the foundational thinking is not mine. Second, defensive value for the reader. When a critic argues that Personal versus Operational is a simplification of automation theory, the answer is yes, deliberately. The simplification is the point.</p><div><hr></div><h2>The four confusions the lens prevents</h2><p>Four common confusions that the lens of &#8220;who runs the task&#8221; surfaces and helps avoid.</p><p><strong>The naming confusion.</strong> Person-Led AI does not mean &#8220;AI for personal use&#8221; in the consumer sense. A team of fifty analysts using Claude to write reports is Person-Led AI. A single executive using ChatGPT at home is Person-Led AI. The &#8220;person&#8221; refers to who runs the task: a person, in both cases.</p><p>Reset the term in your head to &#8220;person-led AI&#8221;, and the framework lands faster.</p><p><strong>The dichotomy confusion.</strong> No real deployment is purely one or the other. Most workflows have moments of Person-Led AI and moments of Operational AI woven together.</p><p>The framework is not a label that goes on a system. It is a question that points at where the fulcrum of the work sits. If the fulcrum is the person initiating and finishing, Person-Led AI is the dominant mode. If the fulcrum is the agent initiating and finishing, Operational AI is the dominant mode.</p><p>Most systems sit somewhere on the spectrum, but the dominant mode determines what you have to invest in: training, infrastructure, governance.</p><p><strong>The choice-versus-constraint confusion.</strong> Often &#8220;who runs the task&#8221; is not a strategic decision. It is an artifact of what the vendor exposes, what the team can configure, and what the existing workflow tolerates.</p><p>An organization that bought a copilot product expecting Operational AI returns may discover that the configuration surface only supports Person-Led AI flows. The framework is useful for diagnosing the present (what do you have today?) and for choosing the next twelve months (where do you want the fulcrum to move?). These are different uses, and they should not be conflated.</p><p><strong>The model-capability confusion.</strong> This framework does not describe what models can do. It describes how organizations choose to deploy them.</p><p>As the underlying models become more capable (memory, planning, meta-cognition), the technical bar for Operational AI lowers. The organizational bar does not. Choosing to let an agent run a workflow without you is, and will remain, an organizational decision more than a technological one.</p><p>The framework does not expire when the next model arrives.</p><div><hr></div><h2>A diagnostic you can run this week</h2><p>Before deciding what to do next, audit what you have today. The diagnostic is short, and the questions are deliberately simple.</p><p>Pick one AI deployment that exists in your organization right now. A Copilot rollout, a customer service bot, an analytics agent, anything in production. Then answer five questions about it.</p><p>First, who initiates the task? A person, every time? A trigger, every time? A mix?</p><p>Second, who finishes the task? Does the loop close on the person&#8217;s review, or does the agent close it on its own?</p><p>Third, if the people who currently run this work all left tomorrow, does the value persist? Or does it walk out the door with them?</p><p>Fourth, what does the trigger configuration look like?</p><blockquote><p>If it is &#8220;a person decides to use the tool&#8221;, you have Person-Led AI regardless of what the vendor calls it. If it is &#8220;a configured event makes the system act&#8221;, you have Operational AI regardless of how simple the underlying model is.</p></blockquote><p>Fifth, what is the cost model? Per-seat licensing scales with people, which scales with Person-Led AI. Token usage and infrastructure scales with workload, which scales with Operational AI. The cost shape often reveals what the deployment actually is.</p><p>Run this on three or four deployments. The pattern will become clear.</p><p>Most enterprises discover, when they do this honestly, that the overwhelming majority of what they have is Person-Led AI, that they have been positioning some of it as Operational AI in conversations with leadership, and that the gap between actual and described is contributing to the disappointment in returns.</p><div><hr></div><h2>The honest caveat: why Operational AI fails today</h2><p>I have to be careful here, because I do not want to sell Operational AI as a guaranteed win. The data does not support that claim, and the failures are instructive.</p><p><a href="https://www.gartner.com/en/newsroom/press-releases/2025-06-25-gartner-predicts-over-40-percent-of-agentic-ai-projects-will-be-canceled-by-end-of-2027">Gartner predicts</a> that more than 40% of agentic AI projects will be canceled by the end of 2027, citing escalating costs, unclear business value, and inadequate risk controls.</p><p>According to <a href="https://fortune.com/2025/12/09/harvard-business-review-survey-only-6-percent-companies-trust-ai-agents/">Harvard Business Review survey data</a>, only 6% of companies fully trust AI agents to handle core business processes. ASAPP reports that <a href="https://www.asapp.com/blog/inside-the-ai-agent-failure-era-what-cx-leaders-must-know">88% of customer service AI agent projects never reach full production</a>.</p><p>Klarna, the most public case of aggressive Operational AI deployment in customer service, <a href="https://www.digitalapplied.com/blog/klarna-reverses-ai-layoffs-replacing-700-workers-backfired">reversed course in 2025</a> and began rehiring staff after the cost-driven optimization compromised customer experience.</p><p>These failures are not random. They share a pattern.</p><p>In every case I have looked at, Operational AI failed for the same reason: no one in the organization had the authority to guide the entire transition. Tool selection sat with IT. Workflow redesign sat with operations. Governance sat with legal. Talent sat with HR. Budget sat with finance.</p><p>The agent was deployed, and the surrounding organizational redesign was not.</p><blockquote><p>Operational AI does not fail because the technology fails. It fails because the organization treats it as a tool purchase rather than as a transition.</p></blockquote><p>The cancelled projects, the rolled-back deployments, the trust gaps, all share a structural cause. There was no one whose job it was to own the arc from technology selection through workflow redesign through governance through talent. There was no one with the authority to say no to a deployment that was not ready, or yes to a deployment that needed organizational change before it could land.</p><p>The companies that capture value with Operational AI have one thing in common.</p><p>Someone, with title and authority, owns the transition end to end. Sometimes it is a Chief AI Officer with real scope. Sometimes it is a VP of Operations with the mandate to redesign workflows and the budget to back it. Sometimes it is a small team that sits across functions.</p><p>The role matters less than the authority. What matters is that the transition has a single owner.</p><p>This is the conversation I am most interested in for the months ahead. What does that authority look like? What competencies does it require, on the agent side and on the human side? What does the organizational map of an Operational AI deployment actually look like, and how does it get built?</p><p>I will return to these questions in upcoming posts.</p><div><hr></div><h2>A position I should state clearly</h2><p>Before closing, an explicit positioning that affects what comes next in this newsletter.</p><p>Person-Led AI has real, durable value. It is not going away. The 86% of horizontal AI spending captured by copilots is not wasted, and the productivity gains are real. If you are running a Person-Led AI deployment well, keep doing it.</p><p>But the frontier of enterprise value, the place where AI compounds rather than evaporates, is Operational AI.</p><p>That is where I will be focusing my analysis in the months ahead. The framework I am building, the case studies I will discuss, the diagnostic tools I will share, will be oriented toward helping enterprise leaders move from Personal-dominated portfolios to portfolios where Operational AI captures meaningful value.</p><p>This is a position, and I am stating it openly because the alternative is leaving it implicit, which would be worse.</p><p>Readers who want pure neutrality on the Personal versus Operational question may find the focus of upcoming posts uncomfortable. Readers who want a lens for moving capital and attention toward the part of AI that compounds will find this useful.</p><div><hr></div><h2>What to do this week</h2><p>Three actions, in order of leverage.</p><p><strong>First, run the diagnostic above on your three largest AI deployments.</strong> Discover, honestly, what you actually have. The pattern that emerges is more useful than any individual finding. If three out of three turn out to be Person-Led AI, you have a portfolio question. If one is Operational and two are Personal, you have a sequencing question. If you cannot tell which is which, you have a clarity question that needs to be solved before any other decision.</p><p><strong>Second, identify one recurring task in your organization that runs today on human initiation, and could plausibly run on a configured trigger instead.</strong> Monthly reports, weekly digests, ticket routing, compliance summaries, anything that happens at predictable cadence. Do not commit to building it yet. Just identify it, and notice what stops the trigger from being configured today. The blocker is usually not technological. It is usually that nobody owns the redesign.</p><p><strong>Third, identify who in your organization has the authority to guide a transition from Personal-dominant AI to Operational-dominant AI.</strong> This is more specific than &#8220;who owns AI strategy&#8221;.</p><p>The owner of the transition needs authority that crosses four domains at once. Technology selection (which platforms, which vendors). Workflow redesign (which processes get reshaped, which get retired). Governance and risk (what the agent is and is not allowed to do, who answers when it fails). Talent (who gets hired, who gets retrained, who exits, what the new roles look like).</p><p>Most organizations have someone who owns one or two of these domains. Almost none have someone with formal authority across all four.</p><p>The transition requires that person, or a small team that collectively holds those four authorities. Without it, the technology arrives, the workflow does not adapt, the governance lags, the people are caught off guard, and the deployment quietly drifts into Person-Led AI dressed up as something else.</p><p>If you cannot name the person in your organization who has this authority, that is your first work.</p><blockquote><p>Tools come later. The owner comes first.</p></blockquote><p>The most expensive errors in enterprise AI in 2026 are happening because the organizations buying Operational AI did not have anyone capable of guiding the transition.</p><p>The map of categories from last week, and the lens of who runs the task from this week, are tools to make the conversation possible. The work, as always, is in the rooms where the conversation gets had.</p><div><hr></div><h2>Sources &amp; Further Reading</h2><ol><li><p><a href="https://ieeexplore.ieee.org/document/844354">Parasuraman and Sheridan - A Model for Types and Levels of Human Interaction with Automation</a> (2000) - the foundational reference for human-in-the-loop and human-on-the-loop autonomy frameworks</p></li><li><p><a href="https://menlovc.com/perspective/2025-the-state-of-generative-ai-in-the-enterprise/">Menlo Ventures - 2025 State of Generative AI in the Enterprise</a> - copilots at 86% of horizontal AI spend, agent platforms at 10%, 160,000 organizations creating 400,000 custom agents in Q1 FY2026</p></li><li><p><a href="https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai">McKinsey - The State of AI 2025</a> - 6% high performers attributing 5% or more EBIT to AI, workflow redesign as multiplier</p></li><li><p><a href="https://www.gartner.com/en/newsroom/press-releases/2025-08-26-gartner-predicts-40-percent-of-enterprise-apps-will-feature-task-specific-ai-agents-by-2026-up-from-less-than-5-percent-in-2025">Gartner - 40% of Enterprise Applications Will Feature AI Agents by 2026</a> - the steepest enterprise adoption curve on record</p></li><li><p><a href="https://www.gartner.com/en/newsroom/press-releases/2025-06-25-gartner-predicts-over-40-percent-of-agentic-ai-projects-will-be-canceled-by-end-of-2027">Gartner - Over 40% of Agentic AI Projects Will Be Canceled by End of 2027</a> - cancellation forecast, root causes</p></li><li><p><a href="https://www.gartner.com/en/articles/hype-cycle-for-agentic-ai">Gartner - Hype Cycle for Agentic AI</a> - market sizing, agent washing, peak of inflated expectations</p></li><li><p><a href="https://www.innovadeltech.com/salesforce-agentforce-implementation-guide/">Salesforce Agentforce - Implementation Field Data</a> - 40-60% deflection rates in production</p></li><li><p><a href="https://github.blog/news-insights/product-news/github-copilot-meet-the-new-coding-agent/">GitHub - Copilot Coding Agent</a> - autonomous code generation, 30-70% productivity gains</p></li><li><p><a href="https://www.itprotoday.com/it-operations/datadog-dash-2025-autonomous-agents-transform-it-observability-operations">Datadog - Bits AI Autonomous Agents</a> - autonomous SRE, root cause analysis under one minute</p></li><li><p><a href="https://aws.amazon.com/blogs/awsmarketplace/agentic-ai-solutions-in-financial-services/">AWS - Agentic AI in Financial Services</a> - 53% of financial institutions running AI agents in production, $3.50 ROI per $1 invested</p></li><li><p><a href="https://fortune.com/2025/12/09/harvard-business-review-survey-only-6-percent-companies-trust-ai-agents/">Harvard Business Review - Only 6% of Companies Fully Trust AI Agents</a> - trust gap data</p></li><li><p><a href="https://www.asapp.com/blog/inside-the-ai-agent-failure-era-what-cx-leaders-must-know">ASAPP - Inside the AI Agent Failure Era</a> - 88% of customer service AI agent projects never reach full production</p></li><li><p><a href="https://www.digitalapplied.com/blog/klarna-reverses-ai-layoffs-replacing-700-workers-backfired">Klarna - The Reversal</a> - the canonical Operational AI rollback case</p></li><li><p><a href="https://learnexperts.ai/blog/linkedin-workplace-learning-report/">LinkedIn - Workplace Learning Report 2025</a> - AI literacy as the number one rising skill</p></li></ol><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Css3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb73ad1fb-011e-4364-8112-e9fb2e6ee5b8_954x946.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Css3!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb73ad1fb-011e-4364-8112-e9fb2e6ee5b8_954x946.png 424w, https://substackcdn.com/image/fetch/$s_!Css3!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb73ad1fb-011e-4364-8112-e9fb2e6ee5b8_954x946.png 848w, https://substackcdn.com/image/fetch/$s_!Css3!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb73ad1fb-011e-4364-8112-e9fb2e6ee5b8_954x946.png 1272w, https://substackcdn.com/image/fetch/$s_!Css3!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb73ad1fb-011e-4364-8112-e9fb2e6ee5b8_954x946.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Css3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb73ad1fb-011e-4364-8112-e9fb2e6ee5b8_954x946.png" width="954" height="946" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b73ad1fb-011e-4364-8112-e9fb2e6ee5b8_954x946.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:946,&quot;width&quot;:954,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:776351,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://marcosnewsletter.substack.com/i/196538566?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb73ad1fb-011e-4364-8112-e9fb2e6ee5b8_954x946.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Css3!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb73ad1fb-011e-4364-8112-e9fb2e6ee5b8_954x946.png 424w, https://substackcdn.com/image/fetch/$s_!Css3!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb73ad1fb-011e-4364-8112-e9fb2e6ee5b8_954x946.png 848w, https://substackcdn.com/image/fetch/$s_!Css3!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb73ad1fb-011e-4364-8112-e9fb2e6ee5b8_954x946.png 1272w, https://substackcdn.com/image/fetch/$s_!Css3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb73ad1fb-011e-4364-8112-e9fb2e6ee5b8_954x946.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"></div></div></a></figure></div><p></p>]]></content:encoded></item><item><title><![CDATA[Most AI Agents Are Not AI Agents]]></title><description><![CDATA[Four products labeled "AI Agent". Only two actually are. Here's the map, and the distinction that determines whether your AI investment compounds.]]></description><link>https://marcosnewsletter.substack.com/p/most-ai-agents-are-not-ai-agents</link><guid isPermaLink="false">https://marcosnewsletter.substack.com/p/most-ai-agents-are-not-ai-agents</guid><dc:creator><![CDATA[Marco]]></dc:creator><pubDate>Tue, 05 May 2026 12:56:49 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!5fZT!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8991220c-3a99-439e-bc57-c034720de56e_1729x910.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!5fZT!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8991220c-3a99-439e-bc57-c034720de56e_1729x910.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!5fZT!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8991220c-3a99-439e-bc57-c034720de56e_1729x910.png 424w, https://substackcdn.com/image/fetch/$s_!5fZT!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8991220c-3a99-439e-bc57-c034720de56e_1729x910.png 848w, https://substackcdn.com/image/fetch/$s_!5fZT!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8991220c-3a99-439e-bc57-c034720de56e_1729x910.png 1272w, https://substackcdn.com/image/fetch/$s_!5fZT!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8991220c-3a99-439e-bc57-c034720de56e_1729x910.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!5fZT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8991220c-3a99-439e-bc57-c034720de56e_1729x910.png" width="1456" height="766" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8991220c-3a99-439e-bc57-c034720de56e_1729x910.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:766,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2805938,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://marcosnewsletter.substack.com/i/196088212?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8991220c-3a99-439e-bc57-c034720de56e_1729x910.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!5fZT!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8991220c-3a99-439e-bc57-c034720de56e_1729x910.png 424w, https://substackcdn.com/image/fetch/$s_!5fZT!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8991220c-3a99-439e-bc57-c034720de56e_1729x910.png 848w, https://substackcdn.com/image/fetch/$s_!5fZT!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8991220c-3a99-439e-bc57-c034720de56e_1729x910.png 1272w, https://substackcdn.com/image/fetch/$s_!5fZT!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8991220c-3a99-439e-bc57-c034720de56e_1729x910.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>For the past year I have watched the term &#8220;AI agent&#8221; do something unusual in the enterprise vocabulary. It has become the most common word in AI strategy meetings and, simultaneously, the least well-defined. Different people use it to mean different things, often within the same conversation, without realizing they are not talking about the same product, the same investment, or the same risk profile.</p><p>This is not a vocabulary problem. It is a budget problem. When the term means four different things, the request for funding you approve under one definition becomes the deployment that fails under another. The vendor who promised category four delivers category one. The internal team funded to build category two ends up trying to build category four because the executive sponsor expected something else. The post-mortem blames execution. The actual failure was a definition.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://marcosnewsletter.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Marco's Substack is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>What follows is the map I built for myself over the past eighteen months as I worked across enterprise AI deployments. It separates the four meanings of &#8220;AI agent&#8221; that circulate today, places real vendors and products in each category, and then introduces a strategic distinction (Personal AI versus Operational AI) that determines whether the investment compounds or evaporates. I have never seen the four categories laid out together in a way that survives contact with a real procurement decision. This is my attempt.</p><div><hr></div><h2>The four categories</h2><p>The categories are ordered by autonomy. Each one is a real, deployable system today. Each one has serious vendors selling it. Each one has appropriate use cases. The error is not in choosing a category. The error is in not knowing which category you are choosing.</p><h3>Category 1: LLM Chatbot</h3><p>The first and most basic meaning of &#8220;AI agent&#8221; is the conversational interface to a large language model. Type a question, get an answer. The system has no persistent state between sessions, no access to external tools, no ability to act in the world beyond producing text.</p><p>Examples sit in plain view: ChatGPT in a browser, Claude.ai, Gemini in its consumer interface, the customer service chatbot on a retailer&#8217;s website. When a vendor demos &#8220;our AI agent&#8221; and shows a chat window where a user types a question and receives an answer, this is the category being shown.</p><p>The technical literature is consistent on what this is and what it is not. <a href="https://lilianweng.github.io/posts/2023-06-23-agent/">Lilian Weng</a>, formerly head of safety systems at OpenAI and now at Thinking Machines Lab, defined a true agent in 2023 as &#8220;LLM + memory + planning + tool use&#8221;. A pure chatbot has the LLM and nothing else. It is a powerful system. It is not, by Weng&#8217;s definition or anyone else&#8217;s, an agent in the technical sense.</p><p>The category is real and useful. It accelerates research, drafting, ideation, customer self-service for simple inquiries. The category is not, however, what most enterprise sponsors imagine when they fund &#8220;agentic AI&#8221;. And this confusion is where the cost begins.</p><h3>Category 2: AI Assistant or Copilot</h3><p>The second meaning is the AI assistant or copilot, where the system has access to tools (read your email, search your documents, draft a response) but operates under continuous human control. The user remains the decision-maker. The AI proposes, the user disposes.</p><p>Microsoft&#8217;s own documentation draws the line clearly. <a href="https://www.microsoft.com/en-us/microsoft-copilot/copilot-101/copilot-ai-agents">Microsoft says of Copilot</a> that it is &#8220;your personal, private assistant that works solely for you, enhancing your capabilities&#8221;, and that it &#8220;can only see files, emails, and data that the logged-in user has permission to access&#8221; and &#8220;cannot take action without the user&#8217;s input&#8221;. GitHub Copilot for code, Microsoft 365 Copilot for office work, Cursor for development, Notion AI for documents: all four are category 2 systems. The human remains in the loop on every meaningful action.</p><p>Microsoft itself contrasts this with category 3 or 4 systems, which they call agents and describe as &#8220;expert systems that operate autonomously on behalf of a process or company&#8221;. The vendor that built the largest copilot product on the market is the same vendor telling buyers, in their own marketing copy, that copilot is not an agent. Few buyers read carefully enough to notice.</p><p>The category is enormously useful. The deployment risk is low because the human catches errors. The productivity gains are real and measurable at the individual level. What category 2 does not do, and what no honest vendor claims it does, is operate processes without human supervision. It makes one person faster. It does not make a function autonomous.</p><h3>Category 3: Agentic Workflow</h3><p>The third category is where the technical definitions start to bite. An agentic workflow is a system in which an LLM and tools are orchestrated through a predefined sequence of steps, where the LLM may decide which step to take next within the boundaries of that sequence, but where the overall workflow is engineered in advance.</p><p><a href="https://www.anthropic.com/research/building-effective-agents">Anthropic, in their December 2024 guide to building effective agents</a>, draws the cleanest distinction in the literature. They write: &#8220;Workflows are systems where LLMs and tools are orchestrated through predefined code paths.&#8221; The system can be sophisticated. It can call external APIs, query databases, send emails, update CRM records. What makes it a workflow rather than a true agent is that the path is determined by the engineer, not by the model.</p><p>This is, in 2026, the category that most vendors are actually selling when they say &#8220;agent&#8221;. Salesforce Agentforce, in nearly all its production deployments, is a category 3 system. <a href="https://www.salesforce.com/agentforce/metrics/">Salesforce&#8217;s own metrics</a> report that Agentforce serves 18,500 enterprise customers and processes more than three billion automated workflows monthly. Note the word: workflows. The number is large because the underlying systems are well-bounded, predictable, and engineered. They do not decide what to do. They execute what was designed.</p><p>The category is the workhorse of enterprise AI in 2026. It is also the category most often misrepresented to executives, who hear &#8220;agent&#8221; and imagine category 4, then approve budgets sized for category 4, then receive category 3, then declare the project a partial success because expectations were set against the wrong reference.</p><h3>Category 4: Autonomous Agent</h3><p>The fourth category is the autonomous agent in the strict technical sense. An LLM dynamically directs its own processes and tool usage. The system decides what step to take next based on the result of the previous step. It plans, it self-corrects, it can pursue a goal across many steps without being told the path in advance.</p><p><a href="https://www.anthropic.com/research/building-effective-agents">Anthropic continues the same passage</a>: &#8220;Agents are systems where LLMs dynamically direct their own processes and tool usage, maintaining control over how they accomplish tasks.&#8221; <a href="https://www.deeplearning.ai/courses/agentic-ai/">Andrew Ng&#8217;s four agentic design patterns</a> (Reflection, Tool Use, Planning, and Multi-Agent Collaboration) all sit in this category. The system reflects on its own output, decides to use a tool, plans multiple steps ahead, or coordinates with other specialized agents to complete a task.</p><p>This is the category that gets the magazine covers. AutoGPT in 2023, OpenAI Operator in 2025, the multi-agent research systems built on top of LangGraph and CrewAI, the experimental autonomous coding agents at the frontier labs. The systems in this category are real. They also fail more often, cost more to operate, are harder to govern, and require more sophisticated infrastructure than any of the other three.</p><p><a href="https://www.gartner.com/en/articles/hype-cycle-for-agentic-ai">Gartner&#8217;s 2026 Hype Cycle for Agentic AI</a> places autonomous agents at the Peak of Inflated Expectations and predicts that <a href="https://www.gartner.com/en/newsroom/press-releases/2025-06-25-gartner-predicts-over-40-percent-of-agentic-ai-projects-will-be-canceled-by-end-of-2027">over 40% of agentic AI projects will be canceled by the end of 2027</a>, citing escalating costs, unclear business value, and inadequate risk controls. The category is where the most ambitious deployments live, and also where the highest failure rates concentrate. Both facts are true at the same time.</p><p>Anthropic&#8217;s own guidance to teams considering building in this category is worth quoting directly. They recommend &#8220;finding the simplest solution possible, and only increasing complexity when needed. This might mean not building agentic systems at all.&#8221; A frontier lab is telling the market: most of you do not need this category. Most of you should be building category 2 or category 3.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://marcosnewsletter.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">If you find this useful, consider subscribing. I publish data-driven breakdowns on enterprise AI maturity every Tuesday, no hype, no vendor pitches.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><h2>The strategic distinction: Personal AI versus Operational AI</h2><p>The four categories above are the technical map. Useful, but not yet sufficient. A technical map tells you what each category is. It does not tell you what each category does for your organization, nor where the value accumulates after deployment. For that you need a second layer of distinction, which is where this article shifts from descriptive to strategic.</p><p>The four categories collapse cleanly into two strategic types.</p><p>Categories 1 and 2 are what I have been calling Personal AI. They amplify an individual. The output is a more productive person. The value, when realized, sits inside the individual: their faster drafting, their better research, their improved code, their accelerated decisions. When that individual leaves the company, the value leaves with them, because the value was never codified into the organization&#8217;s operating infrastructure.</p><p>Categories 3 and 4 are Operational AI. They run a process. The output is a workflow that completes without the human bottleneck that previously gated it. The value, when realized, sits inside the infrastructure of the process: the SLA that holds when volume triples, the response time that stays low at midnight, the cost-per-transaction that drops because no human is involved in the routine cases. When individuals leave, the value remains, because the value was codified into how the work runs.</p><p>I introduced this distinction in the first article of this series, <a href="https://claude.ai/local_sessions/link-al-post-01">From ChatGPT to Operational AI</a>, and the more deployments I observe the more central it becomes to predicting which projects compound and which evaporate. Other thinkers have proposed parallel distinctions. <a href="https://blogs.infosys.com/emerging-technology-solutions/artificial-intelligence/why-ai-in-the-enterprise-is-not-enterprise-ai-the-operating-model-difference-that-most-organizations-miss.html">Infosys, for example, distinguishes &#8220;AI in the Enterprise&#8221; from &#8220;Enterprise AI&#8221;</a> through a lens of governance and decision authority. The framing I use here, developed independently, distinguishes by where value accumulates rather than by who has decision rights. The two lenses are complementary, but they produce different recommendations on where to invest and how to measure return.</p><p>The single most damaging confusion in enterprise AI in 2026 is the mismatch between purchase and expectation across these two types. Companies buy Personal AI (categories 1 and 2) and expect Operational AI returns (the kind of P&amp;L impact that comes from categories 3 and 4). The numbers cannot work. They were never going to work. The investment was sized for one outcome and the technology is engineered for another.</p><p><a href="https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai">McKinsey&#8217;s State of AI 2025</a> puts a sharp number on this. Only 6% of organizations qualify as high performers generating 5% or more EBIT impact from AI initiatives. The high performers are 3.6 times more likely than others to pursue transformational change, and 3 times more likely to have fundamentally redesigned individual workflows when deploying AI. The pattern is unambiguous. Organizations that capture meaningful value are operating in the Operational AI lane, with the appropriate categorical investment. The 94% that do not are, in significant part, organizations that bought Personal AI and waited for Operational AI returns. The wait is structural. It will not arrive.</p><div><hr></div><h2>The next threshold: function-level automation</h2><p>There is a further step beyond Operational AI that is starting to appear in the most aggressive deployments, and naming it correctly matters because the gap between Operational AI and this next state is a change of nature, not of scale.</p><p>I call it function-level automation. The pattern: a category 4 system, or a coordinated set of category 3 and 4 systems, is deployed against an entire business function (customer service, marketing analytics, parts of finance), and that function operates without dedicated human headcount on routine operations. Humans remain involved in strategic supervision, exception handling, escalation, and policy. Routine execution runs autonomously.</p><p>This is not the same as Operational AI applied to a single workflow. Function-level automation is the threshold at which an organization stops &#8220;using AI&#8221; and starts being &#8220;partially executed by AI&#8221;. A change of category, with implications for governance, risk, headcount planning, and competitive positioning that are different from anything Operational AI raised.</p><p>Two cases worth studying together, because they are opposite extremes and both are real.</p><p>The aspirational case is Shopify. In a <a href="https://www.cnbc.com/2025/04/07/shopify-ceo-prove-ai-cant-do-jobs-before-asking-for-more-headcount.html">March 2025 internal memo</a> that became widely public, CEO Tobi Lutke wrote that &#8220;before asking for more headcount and resources, teams must demonstrate why they cannot get what they want done using AI&#8221;, and that AI usage would become &#8220;a baseline expectation&#8221; embedded in performance reviews. The policy was supported by behavior: between 2022 and 2024, <a href="https://fortune.com/2025/04/08/shopify-ceo-ai-automation-no-new-hires-tech-jobs/">Shopify&#8217;s headcount declined from 11,600 to 8,100</a> while the company maintained roughly 21% annual growth. Function-level automation, deliberately pursued, with measurable financial and operational consequences.</p><p>The cautionary case is Klarna. Between 2022 and 2024, Klarna eliminated approximately 700 customer service positions and replaced them with an AI assistant developed in partnership with OpenAI. CEO Sebastian Siemiatkowski <a href="https://www.entrepreneur.com/business-news/klarna-ceo-reverses-course-by-hiring-more-humans-not-ai/491396">told Bloomberg in 2025 that the company was reversing course</a>, citing increased customer complaints, lower satisfaction ratings, and persistent frustration with &#8220;generic, repetitive, and insufficiently nuanced&#8221; replies. His diagnosis of the failure: &#8220;cost was a too predominant evaluation factor&#8221;. Klarna&#8217;s new model is hybrid. AI handles basic inquiries, humans handle anything requiring empathy, discretion, or escalation. Function-level automation attempted, partially reversed, with the lesson absorbed publicly.</p><p>The two cases are bookends. Shopify is the proof that function-level automation can work at scale when the function is amenable. Klarna is the proof that the same approach, applied to a function where empathy and judgment are part of the value, breaks. The lesson is not that one company succeeded and another failed. The lesson is that function-level automation is a categorical decision, not a degree of Operational AI deployment. The questions it raises (which functions, with what governance, accepting which downsides) are different in kind from the questions Operational AI raises.</p><p>The further implication, and this is the part most enterprise leaders have not yet absorbed: function-level automation will not arrive uniformly across the economy. It will arrive function-by-function, company-by-company, with significant variance based on whether the function can be specified well enough to be amplified. The companies who reach the threshold thoughtfully will compound advantages. The companies who reach it accidentally, like Klarna did, will absorb expensive corrections.</p><div><hr></div><h2>The costly confusions</h2><p>Three patterns of mismatch dominate the failed deployments I have observed. Each follows from confusing one category with another. Each is preventable with the map.</p><p>The first mismatch is buying category 2 (Copilot) and expecting category 3 or 4 (Operational AI) returns. A CFO approves Microsoft 365 Copilot licenses for 500 employees, expecting a measurable productivity dividend that translates to headcount efficiency or revenue impact. At 12 months, individuals are faster but processes look the same. Personal AI was deployed. Operational AI returns were expected. The mismatch is structural, not executional, and no amount of &#8220;change management&#8221; will close it.</p><p>The second mismatch is buying a category labeled &#8220;agentic&#8221; from a vendor that is selling category 1 or 2 with new branding. <a href="https://www.gartner.com/en/articles/hype-cycle-for-agentic-ai">Gartner has named this practice &#8220;agent washing&#8221;</a> and estimates that, of the thousands of vendors claiming agentic capabilities, only roughly 130 actually deliver autonomous, goal-pursuing systems. The other 95% are existing products with new marketing. The buyer pays category 4 prices, receives category 1 or 2 functionality, and discovers the gap only after deployment. The defense is procurement discipline: ask the vendor explicitly which category they sit in, and ask them to demonstrate dynamic decision-making rather than predefined workflows.</p><p>The third mismatch is over-engineering: deploying category 4 (autonomous agent) on a problem that category 2 or 3 would solve more cheaply and reliably. This pattern is most common when an internal technical team is excited about new capabilities. Multi-agent systems built for tasks that needed a copilot. Autonomous agents wired up for workflows that were already well-specified. The result is high complexity, high latency, high operational cost, and no incremental benefit over a simpler design. Anthropic&#8217;s own guidance, again worth quoting: &#8220;find the simplest solution possible, and only increase complexity when needed&#8221;. The discipline of choosing the right category is at least as important as choosing the right vendor.</p><div><hr></div><h2>Using the map next week</h2><p>Three actions you can take in the next seven days to put the map to work.</p><p>First, classify the next AI proposal that lands on your desk before reading anything else. Which of the four technical categories is it? Personal AI or Operational AI? Is it function-level, or does it touch only a single workflow? The classification will take five minutes and will reframe every other question you ask about the proposal.</p><p>Second, ask your current AI vendors which category they sit in, and ask them to demonstrate it rather than describe it. A vendor selling category 4 should be able to show you the model dynamically choosing tools and revising plans. A vendor selling category 3 should show you the workflow specification and admit that the path is engineered. A vendor selling category 1 or 2 should not be marketed as &#8220;agentic&#8221;. The honest answers will tell you who to trust.</p><p>Third, audit your existing AI deployments against the Personal versus Operational distinction. How much of your current AI spend is Personal AI? How much is Operational AI? What returns is each generating, and against what expectations were they originally funded? Most organizations discover, when they run this audit honestly, that they are heavily over-invested in Personal AI and under-invested in Operational AI, and that the return profile they were promised cannot be achieved with the current category mix.</p><p>The map does not solve the problem. It surfaces it. From there, the work is choosing where to invest next, which vendors to trust, and what conversations to have inside the organization. But none of that work is possible without first agreeing on what the words mean. The most expensive errors in enterprise AI in 2026 are happening because the people in the room are using the same word to describe different products.</p><p>The work of clarifying terms is unglamorous. It is also, in my observation across the last year of deployments, the most reliable predictor of which AI investments compound. Better thinking, in front of better technology, produces compounding value. Imprecise thinking, in front of any technology, produces fluffy returns. The category names are a small contribution to better thinking. Use them, refine them, push back on them. But do not enter the next AI conversation without them.</p><div><hr></div><h2>Sources &amp; Further Reading</h2><ol><li><p><a href="https://www.anthropic.com/research/building-effective-agents">Anthropic - Building Effective Agents</a> (December 2024, updated 2025) - the canonical distinction between workflows and agents</p></li><li><p><a href="https://lilianweng.github.io/posts/2023-06-23-agent/">Lilian Weng - LLM Powered Autonomous Agents</a> (June 2023) - the foundational definition Agent = LLM + memory + planning + tool use</p></li><li><p><a href="https://www.deeplearning.ai/courses/agentic-ai/">Andrew Ng - DeepLearning.AI Agentic Design Patterns</a> - the four design patterns (Reflection, Tool Use, Planning, Multi-Agent Collaboration)</p></li><li><p><a href="https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai">McKinsey - The State of AI 2025: Agents, Innovation, and Transformation</a> (November 2025) - 6% high performers, workflow redesign as 2.8x EBIT multiplier, scaling agent statistics</p></li><li><p><a href="https://www.gartner.com/en/articles/hype-cycle-for-agentic-ai">Gartner - 2026 Hype Cycle for Agentic AI</a> - agent washing, Peak of Inflated Expectations, deployment statistics</p></li><li><p><a href="https://www.gartner.com/en/newsroom/press-releases/2025-06-25-gartner-predicts-over-40-percent-of-agentic-ai-projects-will-be-canceled-by-end-of-2027">Gartner - Over 40% of Agentic AI Projects Will Be Canceled by End of 2027</a> (June 2025) - failure rate prediction and causes</p></li><li><p><a href="https://www.microsoft.com/en-us/microsoft-copilot/copilot-101/copilot-ai-agents">Microsoft - Copilot and AI Agents</a> - vendor&#8217;s own definitions distinguishing Copilot from autonomous agents</p></li><li><p><a href="https://www.entrepreneur.com/business-news/klarna-ceo-reverses-course-by-hiring-more-humans-not-ai/491396">Klarna - CEO Reverses Course on AI Customer Service</a> (Entrepreneur, 2025) - the canonical reverse case</p></li><li><p><a href="https://www.cnbc.com/2025/04/07/shopify-ceo-prove-ai-cant-do-jobs-before-asking-for-more-headcount.html">Shopify - CEO&#8217;s AI Memo on Headcount Policy</a> (CNBC, April 2025) - the aspirational function-level automation case</p></li><li><p><a href="https://www.salesforce.com/agentforce/metrics/">Salesforce - Agentforce Production Metrics</a> - the largest production deployment of category 3 systems in 2026</p></li><li><p><a href="https://blogs.infosys.com/emerging-technology-solutions/artificial-intelligence/why-ai-in-the-enterprise-is-not-enterprise-ai-the-operating-model-difference-that-most-organizations-miss.html">Infosys - Why &#8220;AI in the Enterprise&#8221; Is Not Enterprise AI</a> - parallel framework distinguishing assistive intelligence from decision-participating intelligence</p></li></ol><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://marcosnewsletter.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Marco's Substack is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Stop Investing in Better AI. Start Investing in Better Thinking.]]></title><description><![CDATA[Last week I argued that the regulatory frameworks your CISO follows have nothing to say about AI agents. This week: even with perfect governance, most AI investments would still fail. Because the real]]></description><link>https://marcosnewsletter.substack.com/p/stop-investing-in-better-ai-invest-in-better-thinking</link><guid isPermaLink="false">https://marcosnewsletter.substack.com/p/stop-investing-in-better-ai-invest-in-better-thinking</guid><dc:creator><![CDATA[Marco]]></dc:creator><pubDate>Tue, 28 Apr 2026 10:43:46 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!MjjL!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6028dcbc-ba35-42ea-9f6e-c52d670169dd_1729x910.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!MjjL!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6028dcbc-ba35-42ea-9f6e-c52d670169dd_1729x910.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!MjjL!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6028dcbc-ba35-42ea-9f6e-c52d670169dd_1729x910.png 424w, https://substackcdn.com/image/fetch/$s_!MjjL!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6028dcbc-ba35-42ea-9f6e-c52d670169dd_1729x910.png 848w, https://substackcdn.com/image/fetch/$s_!MjjL!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6028dcbc-ba35-42ea-9f6e-c52d670169dd_1729x910.png 1272w, https://substackcdn.com/image/fetch/$s_!MjjL!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6028dcbc-ba35-42ea-9f6e-c52d670169dd_1729x910.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!MjjL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6028dcbc-ba35-42ea-9f6e-c52d670169dd_1729x910.png" width="1456" height="766" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/6028dcbc-ba35-42ea-9f6e-c52d670169dd_1729x910.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:766,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2739254,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://marcosnewsletter.substack.com/i/195599910?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6028dcbc-ba35-42ea-9f6e-c52d670169dd_1729x910.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!MjjL!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6028dcbc-ba35-42ea-9f6e-c52d670169dd_1729x910.png 424w, https://substackcdn.com/image/fetch/$s_!MjjL!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6028dcbc-ba35-42ea-9f6e-c52d670169dd_1729x910.png 848w, https://substackcdn.com/image/fetch/$s_!MjjL!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6028dcbc-ba35-42ea-9f6e-c52d670169dd_1729x910.png 1272w, https://substackcdn.com/image/fetch/$s_!MjjL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F6028dcbc-ba35-42ea-9f6e-c52d670169dd_1729x910.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>For the past several months I have been working through a question that I cannot get out of my head.</p><p>Every credible study on enterprise AI in 2025-2026 reports the same broad pattern: high adoption, low economic impact, rising abandonment rates. I have read these reports carefully. I have traced the citations back to the primary research. I have laid the surveys side by side, looked at sample sizes, compared methodologies. The pattern is not a research artifact. It is the dominant story of enterprise AI right now.</p><p>What confuses me is not the data. It is the response.</p><p>Most of the public conversation around AI in 2026 is still about better models, longer context windows, faster inference, smarter agents. The implicit assumption is that the technology is not yet good enough, and that the next release will close the gap. But the data does not say the technology is not good enough. The data says the technology has been deployed at scale and has failed to compound. Those are different problems. They have different solutions.</p><p>The more I sit with this, the more convinced I become that the framing is upside down. We treat AI value capture as a technology question because the technology is the visible part. We chase model upgrades and new platforms because those are easy to see and easy to buy. The actual variable that determines whether AI investment compounds is upstream of the model, sits inside the organization, and is harder to see because it is intangible.</p><p>This is the argument I want to lay out in this piece: most enterprise AI failures are not failures of technology. They are failures of systematic thinking. The model is not the bottleneck. The clarity that should come before the model is.</p><div><hr></div><h2>The numbers do not describe a technology problem</h2><p>Start with the headline statistic that has been circulating since last summer. MIT&#8217;s Project NANDA, in its July 2025 <em>State of AI in Business 2025</em> report, found that 95% of enterprise AI pilots delivered no measurable P&amp;L impact, despite $30-40 billion in investment over two years.</p><p>That single figure has been quoted in roughly every AI strategy deck I have seen since. What gets quoted less often is the company alongside it.</p><p>S&amp;P Global&#8217;s <em>Voice of the Enterprise</em> survey of 1,006 IT and business leaders, published in March 2025, reported that 42% of companies had abandoned the majority of their AI initiatives before reaching production. The previous year that number was 17%. The abandonment rate had multiplied by 2.5 in twelve months. A separate finding from the same survey: 46% of proofs of concept were scrapped between PoC and production deployment.</p><p>McKinsey&#8217;s <em>State of AI 2025</em> tells a similar story from a different angle. 88% of organizations now use AI in at least one function. Only 39% report any EBIT impact from it. And just 6% qualify as what McKinsey calls &#8220;high performers&#8221; on AI value capture. Adoption is high. Impact is low. The gap is widening, not narrowing.</p><p>BCG, in its September 2025 report <em>The Widening AI Value Gap</em>, classified only 5% of companies as &#8220;future-built&#8221; on AI. Those 5% achieve 1.7x revenue growth and 3.6x three-year shareholder returns compared to peers. The other 95% are split between 35% scaling without winning and 60% with minimal gains.</p><p>Forrester, in its <em>2026 Predictions</em> released in October 2025, predicted that enterprises will defer 25% of planned AI spending to 2027. The reason is not that the technology disappointed. It is that CFOs are losing patience with strategies that cannot show ROI linkage.</p><p>Five independent sources, five different methodologies, one consistent picture. The pattern is structural. It is not a story of immature technology that will mature in another quarter. It is a story of organizations that adopted AI faster than they figured out what to do with it.</p><p>If that were a technology problem, you would expect different vendors, different model generations, different deployment patterns to produce different outcomes. They do not. The failure rate is roughly stable across model families, across no-code and code-first platforms, across cloud providers, across consulting partners. What stays constant across all the variation is the human side of the equation.</p><div><hr></div><h2>What MIT actually concluded (and almost nobody quotes)</h2><p>The 95% number from MIT travels everywhere. The conclusion MIT itself drew from that number travels almost nowhere. Worth reading the report&#8217;s own framing carefully.</p><p>MIT did not describe the gap as a technology gap. It described it as a &#8220;learning gap&#8221; and a problem of &#8220;organizational integration.&#8221; The single most striking finding, buried under the headline number, is that 73% of failed AI projects had no clear executive alignment on what success would look like.</p><p>Read that again. Three out of four failed pilots could not produce, when asked, a written agreement among the executives who had funded them about what would constitute a successful outcome. Not a model accuracy target. Not a budget threshold. A definition of success.</p><p>If you do not have that, no model will save you. A perfect model produces an ambiguous result, and the ambiguous result gets relitigated by stakeholders with different unspoken priors. Eventually the project gets quietly defunded, not because it failed, but because nobody can defend why it succeeded.</p><p>A second finding from the same MIT study reinforces the diagnosis. AI deployments that used external vendor partnerships succeeded 67% of the time. Internal builds succeeded 33% of the time. Half the rate.</p><p>The intuitive read is that vendors are technically better. That read is wrong. The teams building internally have access to the same models, the same tooling, often better data. The difference is that vendor engagements force a contract. Contracts force a written scope. Written scope forces the buyer to articulate what they want before money changes hands. The vendor advantage is not engineering. It is the discipline that the procurement process imposes on the buyer&#8217;s thinking.</p><p>McKinsey&#8217;s <em>State of AI 2025</em> offers a third data point that points the same direction. Organizations that redesigned their workflows to integrate AI achieved significantly more value than those who simply layered AI onto existing processes. McKinsey quantifies this as a 2.8x multiplier on EBIT impact. Workflow redesign, by definition, requires articulating how work currently flows. Most organizations have never done that exercise. AI forces them to, and the ones who actually do it win.</p><p>Three independent findings, three different research teams, one consistent direction: the predictor of success is the upstream clarity of the human side, not any property of the AI itself.</p><div><hr></div><h2>AI is an amplifier, not a tool</h2><p>Here is the mental model that, in my experience, makes the data click into place: AI is not a tool. It is an amplifier.</p><p>A tool does a defined thing. A hammer drives nails. A spreadsheet calculates sums. The output is bounded by the tool&#8217;s design. You can be vague about your intent and still get a usable result, because the tool&#8217;s purpose is narrow enough to compensate.</p><p>An amplifier is different. It takes whatever you put in and makes it louder. If what you put in is signal, you get more signal. If what you put in is noise, you get more noise. The amplifier itself is neutral. The character of the output is set entirely by the input.</p><p>This is what enterprise AI behaves like in 2026. Feed it a workflow you understand precisely, and it will execute that workflow at scale. Feed it a vague brief, and it will produce something that looks like an output but contains nothing actionable. The system does not refuse. It does not error. It produces what I have come to call the &#8220;fluffy&#8221; failure mode: text that has the form of a result but the substance of filler.</p><p>The fluffy failure mode is the mode that should worry executives most, because it is invisible until you read carefully. A spectacular AI failure ends a project. A fluffy failure can continue indefinitely. The dashboard ships. The agent runs. The reports get generated. Nothing in the system signals that the outputs are empty.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!s9YJ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F834d8014-8ffe-4faa-9451-6a6f0f946ab5_1729x910.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!s9YJ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F834d8014-8ffe-4faa-9451-6a6f0f946ab5_1729x910.png 424w, https://substackcdn.com/image/fetch/$s_!s9YJ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F834d8014-8ffe-4faa-9451-6a6f0f946ab5_1729x910.png 848w, https://substackcdn.com/image/fetch/$s_!s9YJ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F834d8014-8ffe-4faa-9451-6a6f0f946ab5_1729x910.png 1272w, https://substackcdn.com/image/fetch/$s_!s9YJ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F834d8014-8ffe-4faa-9451-6a6f0f946ab5_1729x910.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!s9YJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F834d8014-8ffe-4faa-9451-6a6f0f946ab5_1729x910.png" width="1456" height="766" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/834d8014-8ffe-4faa-9451-6a6f0f946ab5_1729x910.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:766,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2956762,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://marcosnewsletter.substack.com/i/195599910?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F834d8014-8ffe-4faa-9451-6a6f0f946ab5_1729x910.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!s9YJ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F834d8014-8ffe-4faa-9451-6a6f0f946ab5_1729x910.png 424w, https://substackcdn.com/image/fetch/$s_!s9YJ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F834d8014-8ffe-4faa-9451-6a6f0f946ab5_1729x910.png 848w, https://substackcdn.com/image/fetch/$s_!s9YJ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F834d8014-8ffe-4faa-9451-6a6f0f946ab5_1729x910.png 1272w, https://substackcdn.com/image/fetch/$s_!s9YJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F834d8014-8ffe-4faa-9451-6a6f0f946ab5_1729x910.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p> Only a human, comparing what they wanted with what they got, can detect the gap. And humans, busy and trusting, often do not.</p><p>The reason this happens is mechanical, not mysterious. Modern language models, agentic or otherwise, are interpretation engines. When given a clear, bounded specification, they execute it. When given an ambiguous brief, they do what they were trained to do: produce plausible-sounding output that matches the surface pattern of the request. Plausible-sounding output is not the same as useful output. It rarely surfaces the friction in your business, because friction is specific and language models reach for the generic.</p><p>Examples are easy to find once you know what to look for. The market analysis report that reads cleanly but tells you nothing about your specific market. The customer support agent that handles 10,000 tickets and escalates the same 50% it always did, because the escalation criteria were never defined precisely enough for the agent to do anything but match the historical pattern. The strategy memo that quotes your industry, your geography, your stakeholders, and somehow ends up indistinguishable from the strategy memo for any other company in your sector.</p><p>These are not technology failures. The model worked. The pipeline ran. The output was produced. What was missing was the human-side specification of what counts as a useful answer in this specific organization, with this specific data, for this specific decision.</p><blockquote><p>If you&#8217;re finding this useful, consider subscribing. I publish data-driven breakdowns on enterprise AI maturity every Tuesday, no hype, no vendor pitches. </p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://marcosnewsletter.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://marcosnewsletter.substack.com/subscribe?"><span>Subscribe now</span></a></p></blockquote><p>There is a related point worth naming. AI does not just amplify your input. It exposes what your input actually was. A vague brief, fed to a junior analyst, comes back with follow-up questions. The analyst asks for clarification, and the act of clarifying surfaces the vagueness for the person who wrote the brief. AI does not ask. It produces. The producer of the brief is left with output but not with the diagnostic moment in which the brief itself gets better.</p><p>This is why I have started to think of AI as a mirror as much as an amplifier. It reflects, with brutal literalness, the quality of thinking that went into the request. The organizations that struggle most with AI are not, in my observation, the ones with the worst data or the smallest budgets. They are the ones whose internal language for what they do has always been imprecise, and who are now meeting a system that cannot translate imprecision into value.</p><div><hr></div><h2>What systematic thinking actually means</h2><p>If &#8220;systematic thinking&#8221; sounds like consulting jargon, that is partly because the consulting industry has overused the phrase. Strip the jargon away and the working definition is concrete.</p><p>Systematic thinking is the ability to externalize three things in writing: the decisions a process contains, the criteria those decisions use, and the reasoning behind those criteria. It is not raw intelligence. It is not strategic vision. It is not even, primarily, a skill in writing. It is the discipline of moving knowledge that lives in someone&#8217;s head into a form that another mind, human or otherwise, can follow.</p><p>Most organizations have very little of it. Most executives discover this only when they try to use AI. The discovery is uncomfortable, because the absence of systematic thinking is not anyone&#8217;s fault. It is the natural state of organizations that grew through human improvisation, where the people who knew how things worked simply did them, and the doing was good enough.</p><p>A simple test. Imagine a smart, capable new hire arrives at your company on Monday morning. They have the right background but no internal knowledge of your specific operation. You assign them a workflow your team owns. Can they, from the documentation that exists, perform that workflow correctly without asking anyone? Or do they need a colleague to fill in the gaps in real time?</p><p>The answer is almost always the second. The colleague is the carrier of the unwritten rules. Improvisation, conversation, judgment, and tribal knowledge fill the space between the formal documentation and the actual work. This is fine for a human-only system, because the colleague is always available. It breaks down the moment you replace the colleague with an AI agent. The agent has the formal documentation. It does not have the colleague.</p><p>Two consequences follow from this.</p><p>The first is that organizations with strong, written process standards get amplified at the level of process. AI in those organizations can take over functions, not just tasks. Workflows can run with limited human intervention because the workflows were specifiable in the first place. This is the path that produces the McKinsey 2.8x multiplier. It is also the path that almost nobody is on.</p><p>The second is that organizations whose work depends on improvisation and tacit knowledge get amplified at the level of person. A skilled employee who learns to use ChatGPT well becomes more productive. The same employee, in the same organization, doing the same work, just faster. The amplification is real, but it does not propagate. It does not become organizational capability. The improvement stays inside the individual.</p><p>Both modes can be valuable. Neither one is a moral position. But executives who fund AI initiatives expecting the first mode and operating in conditions that only support the second will be disappointed forever. The technology cannot give you organizational capability when the organization itself runs on tacit individual capability. It can only give you faster individuals.</p><div><hr></div><h2>A diagnostic question</h2><p>The most useful diagnostic I know for AI readiness is a single question, applied to a specific workflow before any AI investment is committed.</p><p>The question: where does process end, and personal judgment begin?</p><p>Ask this of the team that owns the workflow. There are three honest answers.</p><p>The first answer is &#8220;I can draw the line.&#8221; The team knows which steps are governed by rules and which are reserved for human judgment, and they can articulate the boundary. In this state, AI deployment can be targeted and bounded. The rule-governed parts can be automated. The judgment parts can be supported with information and decision aids. Both layers can be measured separately. This is the rare and valuable state.</p><p>The second answer is &#8220;the line is blurry, it depends.&#8221; The team knows the work but cannot describe precisely when judgment overrides process or vice versa. In this state, AI works at task level but not at workflow level. Individual prompts produce useful output. Attempts to automate the entire workflow fail because the system cannot distinguish between deviation that is wisdom and deviation that is error. This is where most organizations actually live.</p><p>The third answer is &#8220;I have never thought about it that way.&#8221; The team operates entirely on accumulated personal judgment, with no internal model of where process ends. In this state, AI deployment will reliably fail &#8220;fluffy&#8221; until the team is forced to articulate the workflow, at which point the deployment usually becomes unnecessary because the act of articulating already revealed the better path. This is the state of most pilots that quietly die after twelve months.</p><p>The diagnostic is cheap. It takes one meeting. It identifies, before any procurement begins, which workflows are ready for which kind of AI investment, which workflows need clarification work first, and which workflows should not be touched until the team can describe its own work to itself.</p><p>The cost of skipping this diagnostic is the cost of the failed pilot. Multiplied by whatever fraction of your AI portfolio is currently funded against workflows in the second and third state.</p><div><hr></div><h2>What to do Monday morning</h2><p>The next AI investment your organization makes does not need to be a model upgrade, a new platform, or a partnership with a different consulting firm. The highest-leverage thing you can do, before any technology decision, is sit with one team that owns one workflow and write the workflow down.</p><p>Inputs. Decisions. Decision criteria. Outputs. Exceptions. The conditions under which a step is skipped, repeated, or escalated. The signals that determine which path the workflow takes. The downstream consequences of each output.</p><p>If the team can produce that document in a few hours, the workflow is ready for AI. The investment will compound. If the team cannot, the workflow is not ready. Funding an AI deployment against it will join the 95% that delivered no measurable P&amp;L impact.</p><p>Most leaders, faced with this exercise, discover something they did not want to know: their organization has fewer well-defined workflows than they thought, and the ones they do have are not the ones generating the most economic value. The valuable work was being done through skilled improvisation by experienced people. AI cannot amplify what was never made explicit.</p><p>The good news is that the path forward is not technological. It is editorial. Every workflow you can articulate is a workflow you can choose to amplify. Every workflow you cannot articulate is a workflow whose value depends on the specific people doing it today, and is not transferrable to any system, AI or otherwise.</p><p>The 5% of companies BCG calls &#8220;future-built&#8221; are not winning because they have access to better models. The models are commodities. They are winning because they did the unglamorous work of writing down how their business actually runs, and only then asked AI to help.</p><p>The bottleneck has never been the technology. It is, and has always been, the quality of the thinking that goes in front of it. Better AI without better thinking produces faster fluffy. Better thinking, even with off-the-shelf AI, produces compounding value.</p><p>Stop investing in better AI. Start investing in better thinking. The order matters more than the tools.</p><div><hr></div><h2>Sources &amp; Further Reading</h2><ol><li><p><a href="https://mlq.ai/media/quarterly_decks/v0.1_State_of_AI_in_Business_2025_Report.pdf">MIT Project NANDA: The GenAI Divide - State of AI in Business 2025</a> (July 2025)</p></li><li><p><a href="https://www.spglobal.com/market-intelligence/en/news-insights/research/ai-experiences-rapid-adoption-but-with-mixed-outcomes-highlights-from-vote-ai-machine-learning">S&amp;P Global Market Intelligence: Voice of the Enterprise - AI &amp; Machine Learning 2025</a> (March 2025)</p></li><li><p><a href="https://media-publications.bcg.com/The-Widening-AI-Value-Gap-Sept-2025.pdf">BCG: The Widening AI Value Gap</a> (September 2025)</p></li><li><p><a href="https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai">McKinsey: The State of AI in 2025 - Agents, Innovation, and Transformation</a></p></li><li><p><a href="https://www.forrester.com/blogs/predictions-2026-ai-moves-from-hype-to-hard-hat-work/">Forrester: Predictions 2026 - AI Moves From Hype To Hard Hat Work</a> (October 2025)</p></li></ol>]]></content:encoded></item><item><title><![CDATA[105 Days to EU AI Act Enforcement. The Regulators Still Don’t Know What an Agent Is]]></title><description><![CDATA[Last week I argued that AI agents are not employees and need systems-level governance. This week: the systems-level governance guidance doesn&#8217;t exist. Not in NIST. Not in the EU AI Act. Not in ISO 420]]></description><link>https://marcosnewsletter.substack.com/p/ai-agent-governance-vacuum-nist-eu-iso</link><guid isPermaLink="false">https://marcosnewsletter.substack.com/p/ai-agent-governance-vacuum-nist-eu-iso</guid><dc:creator><![CDATA[Marco]]></dc:creator><pubDate>Tue, 21 Apr 2026 12:33:25 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!-6QP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b7493d3-253e-4730-9ff1-0c0675614e2a_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!-6QP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b7493d3-253e-4730-9ff1-0c0675614e2a_1536x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!-6QP!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b7493d3-253e-4730-9ff1-0c0675614e2a_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!-6QP!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b7493d3-253e-4730-9ff1-0c0675614e2a_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!-6QP!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b7493d3-253e-4730-9ff1-0c0675614e2a_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!-6QP!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b7493d3-253e-4730-9ff1-0c0675614e2a_1536x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!-6QP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b7493d3-253e-4730-9ff1-0c0675614e2a_1536x1024.png" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7b7493d3-253e-4730-9ff1-0c0675614e2a_1536x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:3547412,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://marcosnewsletter.substack.com/i/194695731?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b7493d3-253e-4730-9ff1-0c0675614e2a_1536x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!-6QP!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b7493d3-253e-4730-9ff1-0c0675614e2a_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!-6QP!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b7493d3-253e-4730-9ff1-0c0675614e2a_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!-6QP!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b7493d3-253e-4730-9ff1-0c0675614e2a_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!-6QP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7b7493d3-253e-4730-9ff1-0c0675614e2a_1536x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>On August 2, 2026, the European Commission starts enforcing the high-risk provisions of the AI Act. Penalties run up to 35 million euros or 7% of global revenue. That&#8217;s 105 days from today.</p><p>In the same window, Gartner projects that 40% of enterprise applications will include AI agents, up from less than 5% a year ago. Many of those deployments will fall squarely inside the Act&#8217;s high-risk categories: credit scoring, fraud detection, employment screening, essential services.</p><p>Here is what the EU AI Office published on its own FAQ page about how the Act applies to these agents:</p><blockquote><p>&#8220;Developments related to AI agents are recent and fast evolving, and the European Commission&#8217;s regulatory considerations are only preliminary at this stage.&#8221;</p></blockquote><p>That is the enforcement authority, four months before penalties begin, telling enterprises that its considerations are preliminary.</p><p>This is the governance vacuum. And it is not an abstract regulatory problem. It is an operational, legal, and financial risk that is already showing up in boardrooms, insurance renewals, and vendor contracts.</p><div><hr></div><h2>What the three frameworks your CISO already follows actually say about agents</h2><p>Walk into any enterprise security committee and three frameworks get cited: NIST AI RMF, the EU AI Act, ISO/IEC 42001. They are the default language for enterprise AI governance.</p><p>Here is what they say about autonomous agents that chain decisions, call external tools, and take actions with limited human oversight.</p><p><strong>NIST AI RMF 1.0 (January 2023)</strong>: nothing. The framework treats AI as a static system that produces outputs for humans to review. It has no concept of an agent that loops, plans, calls tools, or takes action. The companion GenAI Profile released in July 2024 identifies 12 risks specific to generative AI (confabulation, CBRN information, bias) but does not address agents as a distinct architectural pattern.</p><p>NIST itself acknowledged the gap only two months ago. On February 17, 2026, the NIST Center for AI Standards and Innovation formally launched the AI Agent Standards Initiative. The public comment period on the concept paper closed April 2. Final standards are years away.</p><p><strong>The EU AI Act (entered force August 1, 2024)</strong>: no specific agent category. The Act classifies AI systems as prohibited, high-risk, limited-risk, or minimal-risk. Enterprises must map their agents to these categories themselves. The Act&#8217;s Annex III lists high-risk use cases but says nothing about what makes an autonomous agent performing those tasks different from a classifier or a scoring model.</p><p>Enterprises building agents for credit decisions, insurance underwriting, or recruitment must guess which high-risk obligations apply and how, while the authority enforcing the rules publicly describes its thinking as preliminary.</p><p><strong>ISO/IEC 42001:2023</strong>: a management system standard, not a controls framework. It asks organizations to establish governance structures, risk management processes, and lifecycle controls for AI. It does not define what &#8220;good&#8221; looks like for delegation chains, tool-use permissions, runtime monitoring, autonomy limits, or shutdown procedures for an agent that goes rogue.</p><p>You can be ISO 42001 certified and still have zero specific controls on your agents. The standard certifies that you have a process. It does not tell you what the process should contain.</p><p>Three frameworks, three silences. And the enforcement clock is running.</p><div><hr></div><h2>The sector where this gets expensive first: banking</h2><p>Financial services is where the vacuum hits hardest, because banks have been deploying AI agents faster than almost any other regulated sector.</p><p>HSBC reports a 50% reduction in false positives on fraud detection after moving to agent-based anomaly detection. Citi, UBS, DBS, and ING report 20-40% cost reductions on compliance workflows. JPMorgan Chase runs an enterprise ML platform called OmniAI across fraud, risk, and regulatory operations. McKinsey&#8217;s latest research on financial crime describes agent-led transformations of KYC and AML as already underway.</p><p>Then look at what the bank regulators have written about any of this.</p><p>The United States Office of the Comptroller of the Currency&#8217;s foundational model risk management guidance, SR 11-7, was published on April 4, 2011. It predates modern deep learning. It assumes models are validated periodically, produce scores, and feed into human decisions. It does not contemplate a system that plans, calls external tools, and takes action in a continuous loop.</p><p>Banks now have to reconcile 2011 validation language with 2026 autonomous systems. That reconciliation happens inside each bank, with no consistent supervisory standard.</p><p>The European Banking Authority published a factsheet on November 21, 2025 describing the AI Act&#8217;s implications for banking. Here is the operative sentence:</p><blockquote><p>&#8220;The EBA has not identified any immediate need to introduce any new or review existing EBA Guidelines.&#8221;</p></blockquote><p>The supervisor of European banks, writing four months before high-risk enforcement begins, announced it would not update its guidelines. Banks are on their own.</p><p>This is not a story about regulators being slow. It is a story about regulators being candid that they do not yet know what to say.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://marcosnewsletter.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">If you&#8217;re finding this useful, consider subscribing. I publish data-driven breakdowns on enterprise AI maturity every Tuesday, no hype, no vendor pitches.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><h2>Three risks the vacuum is creating right now</h2><p>The absence of regulatory guidance does not mean the absence of risk. It means the risk is moving to places enterprise boards rarely look until it is too late.</p><p><strong>Risk 1: Insurance is tightening faster than governance is maturing.</strong></p><p>Cyber carriers are adding AI endorsements to clarify what is covered and what is not. Professional liability (E&amp;O) policies are adding explicit exclusions for losses caused by AI-driven decision-making, algorithmic non-compliance, and autonomous system failures. ISACA&#8217;s 2025 analysis on AI and cyber insurance is direct about the trend: coverage is narrowing.</p><p>The practical effect is that governance maturity is becoming a prerequisite for insurability. If an enterprise cannot demonstrate a governance program that covers its agents, some carriers will quietly exclude agent-driven losses, raise premiums, or decline to renew. This is happening now, in renewal cycles across 2026.</p><p><strong>Risk 2: Vendor contracts require compliance with frameworks that do not cover agents.</strong></p><p>Large enterprises increasingly require their vendors to be &#8220;compliant with NIST AI RMF&#8221; or &#8220;aligned with ISO/IEC 42001&#8221; in AI procurement clauses. What this means for an agent-based product is undefined. Vendors sign clauses they cannot fully satisfy because the frameworks they are promising compliance with do not specify controls for the thing they are selling.</p><p>This creates legal exposure on both sides. The buyer cannot verify compliance. The vendor cannot prove it. If the agent causes harm, the contract terms will be litigated against frameworks that were silent on the technology.</p><p><strong>Risk 3: Liability precedent is being set by what exists, not by what fits.</strong></p><p>When an AI agent causes measurable harm in a regulated industry in 2026, the first incident will not wait for NIST to finish its standards initiative or the EU AI Office to issue agent-specific guidance. Courts, regulators, and plaintiffs will reach for whatever governance framework exists.</p><p>In banking, that means SR 11-7 becomes the de facto agent regulation, interpreted by supervisors and courts who have no better option. In healthcare, it means HIPAA and the FDA&#8217;s software-as-medical-device guidance. In employment, it means Title VII and the EEOC&#8217;s algorithmic hiring positions. None of these were designed for agents.</p><p>The precedents being set in these early cases will harden into expectations. Enterprises that cannot show systematic thinking about agent governance, documented in writing, with clear decision rights and boundaries, will find themselves defending against frameworks they did not choose and did not design for.</p><div><hr></div><h2>What is emerging, and why it is not enough yet</h2><p>The vacuum is being filled, slowly, by industry-led frameworks. These are useful. They are not a substitute for regulatory guidance.</p><p><strong>OWASP Top 10 for Agentic Applications (December 2025)</strong> is the first peer-reviewed threat framework specifically for autonomous agents. Developed by over 100 industry experts over eight months, it catalogs risk categories including prompt injection, insecure tool integration, insufficient access controls, unintended autonomy, and inadequate monitoring. It is the closest thing to a shared baseline the industry has.</p><p><strong>MITRE ATLAS</strong> added agent-focused techniques in version 5.4.0 in February 2026, including tool poisoning and privilege escalation in agent workflows. It is a threat model for red-teaming, not a governance standard.</p><p><strong>Cloud Security Alliance</strong> published a research note on April 3, 2026 that explicitly names the agent governance gap and proposes controls. Useful for internal program design, not binding.</p><p><strong>The Agentic AI Foundation</strong> (Block, Anthropic, OpenAI, Google, December 2025) is developing interoperability standards like Model Context Protocol and the Agent2Agent protocol. These address how agents communicate, not how they are governed.</p><p>These are all credible efforts. None of them are regulations. A bank that adopts OWASP Top 10 for Agentic Applications is doing more than the minimum. It is also building on documents that no regulator has endorsed as sufficient. If an enforcement action arrives in August, OWASP compliance will be a defense, not a shield.</p><p>The industry is ahead of the regulators. That is unusual and it will not last. The question is how many enterprises will take harm in the gap before the official frameworks catch up.</p><div><hr></div><h2>The honest counterpoint</h2><p>The vacuum is not equally dangerous everywhere. There is a class of agent deployments where existing governance frameworks, stretched a bit, work fine.</p><p>An internal productivity agent that summarizes documents inside a contained environment, with no external actions, no access to customer data, and no autonomous decisions, does not need regulatory-grade governance. Applying the EU AI Act&#8217;s high-risk machinery to this kind of tool is overkill and will slow useful adoption.</p><p>The vacuum matters for a specific class of agents: those that take actions in the real world, make decisions with financial or legal consequences, operate in regulated industries, or touch customer-facing processes. For those, the absence of guidance is material and current.</p><p>The practical implication is triage. Not every agent needs the same governance treatment. But the triage itself is a governance act, and most enterprises are not doing it systematically.</p><div><hr></div><h2>What to do Monday morning</h2><p>Three things you can do this week, even without waiting for regulators.</p><p><strong>First, classify your agents by risk using the EU AI Act categories as a thinking tool</strong>, even if you are not in scope of EU enforcement. The Act&#8217;s four-tier structure (prohibited, high-risk, limited, minimal) is a useful forcing function. Go through your agent portfolio. Assign each agent to a tier. Document the reasoning in one paragraph per agent. The exercise surfaces the agents where the governance vacuum matters most.</p><p><strong>Second, adopt OWASP Top 10 for Agentic Applications as your working baseline</strong> until official frameworks catch up. It is the most credible industry document. It is free. Your security team can use it as a checklist for every agent in production or in development. When auditors and regulators eventually arrive with their own frameworks, you will have evidence of systematic thinking, not ad-hoc choices.</p><p><strong>Third, document your reasoning in writing for each high-risk agent.</strong> What controls are in place. What decisions the agent can make autonomously. What requires human approval. Who is accountable. How anomalies are detected. How the agent gets shut down. Not because a regulator will read these documents today, but because when one does, enterprises that show systematic reasoning get credit and enterprises that do not become easier targets.</p><p>These three actions are the minimum viable governance for agents in 2026. They will not make you compliant with frameworks that do not yet exist. They will put you in the top decile of enterprises that can defend their decisions when someone eventually asks.</p><p>I am developing a structured approach to this gap called AIMF (AI Maturity Framework), now a registered trademark at the EU Intellectual Property Office. It turns the classification, baseline adoption, and documentation into a repeatable process with scoring and a roadmap. I will share more on it in the coming weeks.</p><p>The regulatory vacuum is going to close. It will close faster than most enterprises expect, and it will close in ways that advantage organizations that were already thinking systematically about agent governance. The window to build that muscle, before the first enforcement action makes governance a fire drill, is measured in months.</p><div><hr></div><p><em>Next week: the real cost of AI agents. Token economics, infrastructure overhead, operational staffing. Most enterprises budget for agents like they budget for ML models. They are off by 3 to 5 times. Here is where the real costs hide.</em></p><div><hr></div><h2>Sources &amp; Further Reading</h2><ol><li><p><a href="https://www.nist.gov/caisi/ai-agent-standards-initiative">NIST AI Agent Standards Initiative</a> (February 2026)</p></li><li><p><a href="https://www.federalregister.gov/documents/2026/01/08/2026-00206/request-for-information-regarding-security-considerations-for-artificial-intelligence-agents">Federal Register RFI on AI Agent Security</a> (January 2026)</p></li><li><p><a href="https://genai.owasp.org/resource/owasp-top-10-for-agentic-applications-for-2026/">OWASP Top 10 for Agentic Applications</a> (December 2025)</p></li><li><p><a href="https://artificialintelligenceact.eu/implementation-timeline/">EU AI Act Implementation Timeline</a></p></li><li><p><a href="https://ai-act-service-desk.ec.europa.eu/en/faq">EU AI Office FAQ</a></p></li><li><p><a href="https://www.eba.europa.eu/sites/default/files/2025-11/d8b999ce-a1d9-4964-9606-971bbc2aaf89/AI%20Act%20implications%20for%20the%20EU%20banking%20sector.pdf">EBA Factsheet on EU AI Act Implications for Banking</a> (November 2025)</p></li><li><p><a href="https://www.gartner.com/en/newsroom/press-releases/2025-08-26-gartner-predicts-40-percent-of-enterprise-apps-will-feature-task-specific-ai-agents-by-2026-up-from-less-than-5-percent-in-2025">Gartner: 40% of Enterprise Apps Will Feature AI Agents by 2026</a> (August 2025)</p></li><li><p><a href="https://www.gartner.com/en/newsroom/press-releases/2025-06-25-gartner-predicts-over-40-percent-of-agentic-ai-projects-will-be-canceled-by-end-of-2027">Gartner: 40% of Agentic AI Projects Will Be Canceled by 2027</a> (June 2025)</p></li><li><p><a href="https://www.mckinsey.com/capabilities/risk-and-resilience/our-insights/how-agentic-ai-can-change-the-way-banks-fight-financial-crime">McKinsey: How Agentic AI Can Change the Way Banks Fight Financial Crime</a> (2025)</p></li><li><p><a href="https://atlas.mitre.org/">MITRE ATLAS Framework</a> (version 5.4.0, February 2026)</p></li><li><p><a href="https://labs.cloudsecurityalliance.org/research/csa-research-note-ai-agent-governance-framework-gap-20260403/">CSA Research Note: AI Agent Governance Framework Gap</a> (April 2026)</p></li><li><p><a href="https://www.isaca.org/resources/news-and-trends/isaca-now-blog/2025/cyber-insurance-in-crisis-with-ai-blind-spots">ISACA: Cyber Insurance in Crisis with AI Blind Spots</a> (2025)</p></li><li><p><a href="https://www.federalreserve.gov/supervisionreg/srletters/sr1107.htm">OCC SR 11-7 Model Risk Management Guidance</a> (2011)</p></li><li><p><a href="https://cloudsecurityalliance.org/press-releases/2025/12/18/csa-and-google-cloud-study-finds-governance-maturity-is-strongest-predictor-of-ai-readiness">CSA and Google Cloud: Governance Maturity as Strongest Predictor of AI Readiness</a> (December 2025)</p></li><li><p><a href="https://www.iso.org/standard/42001">ISO/IEC 42001:2023 AI Management System Standard</a></p></li></ol>]]></content:encoded></item><item><title><![CDATA[AI Agents Are Not Employees. Stop Designing Org Charts for Them.]]></title><description><![CDATA[Your AI Agent Doesn't Need a Performance Review. It Needs a Scaffold. Last week I mapped the pilot graveyard - 95% of AI pilots delivering zero P&L impact. This week: the organizational design mistake]]></description><link>https://marcosnewsletter.substack.com/p/ai-agents-not-employees-org-design</link><guid isPermaLink="false">https://marcosnewsletter.substack.com/p/ai-agents-not-employees-org-design</guid><dc:creator><![CDATA[Marco]]></dc:creator><pubDate>Tue, 14 Apr 2026 10:51:40 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!BTDG!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98eee1ab-dd11-4ee5-b57d-886b281ef850_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!BTDG!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98eee1ab-dd11-4ee5-b57d-886b281ef850_1536x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!BTDG!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98eee1ab-dd11-4ee5-b57d-886b281ef850_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!BTDG!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98eee1ab-dd11-4ee5-b57d-886b281ef850_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!BTDG!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98eee1ab-dd11-4ee5-b57d-886b281ef850_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!BTDG!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98eee1ab-dd11-4ee5-b57d-886b281ef850_1536x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!BTDG!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98eee1ab-dd11-4ee5-b57d-886b281ef850_1536x1024.png" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/98eee1ab-dd11-4ee5-b57d-886b281ef850_1536x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:3458536,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://marcosnewsletter.substack.com/i/193664889?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98eee1ab-dd11-4ee5-b57d-886b281ef850_1536x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!BTDG!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98eee1ab-dd11-4ee5-b57d-886b281ef850_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!BTDG!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98eee1ab-dd11-4ee5-b57d-886b281ef850_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!BTDG!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98eee1ab-dd11-4ee5-b57d-886b281ef850_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!BTDG!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F98eee1ab-dd11-4ee5-b57d-886b281ef850_1536x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Scroll through LinkedIn on any given week and you&#8217;ll see some version of the same image: a traditional org chart, hierarchy intact, with robot icons replacing a few human positions. &#8220;AI Task Agent&#8221; reports to &#8220;AI Supervisor.&#8221; &#8220;Human &amp; AI Teams&#8221; sit in one branch, pure AI teams in another. The CEO stays at the top. The structure stays the same. Only the faces change.</p><p>The posts accompanying these images ask reasonable-sounding questions. How do we manage AI agents? What does performance management look like for them? Do agents &#8220;progress&#8221; in capability the way employees grow in their careers? How do we give them feedback?</p><p>These feel like the right questions. They&#8217;re not. They&#8217;re HR questions applied to systems that aren&#8217;t employees. And that category error - treating agents as workers who slot into existing hierarchies - is about to produce the same failure pattern we just mapped in the <a href="https://marcosnewsletter.substack.com/p/pilot-graveyard-ai-failure-rate?r=e6ccc">pilot graveyard</a>, except at a scale and speed that most organizations aren&#8217;t prepared for.</p><div><hr></div><h2>The hierarchy was designed for human constraints. Agents don&#8217;t have them.</h2><p>The org chart is one of the most durable management inventions of the industrial age. It exists because humans have real cognitive limits: a manager can effectively oversee 5-8 direct reports. Information degrades as it passes through hierarchical layers. Coordination costs increase with every node in the network. Synchronous communication is expensive and slow.</p><p>Every structural choice in a traditional hierarchy - span of control, reporting lines, management layers - is an optimization against these human constraints.</p><p>AI agents have none of them.</p><p>An agent doesn&#8217;t need a manager to prioritize its work. It doesn&#8217;t forget context between conversations (if designed properly). It doesn&#8217;t need to be in the room when a decision is made. It can process input from dozens of sources simultaneously. It can be invoked, configured, composed with other agents, and torn down in seconds.</p><p><a href="https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai">McKinsey&#8217;s research on the agentic organization</a> describes what mature agent deployments actually look like: small teams of 2-5 humans orchestrating 50-100 specialized agents. That&#8217;s not a traditional hierarchy with robot icons. That&#8217;s a fundamentally different organizational architecture where the ratio of humans to agents makes the old reporting structure irrelevant.</p><p>The viral org charts get this exactly backwards. They take a structure optimized for human limitations and swap in entities that don&#8217;t share those limitations. It&#8217;s like redesigning a factory around horses, then replacing the horses with trucks but keeping the same barn layout.</p><div><hr></div><h2>Neither tool nor employee: the category agents actually occupy</h2><p>The deeper problem with the &#8220;agents as employees&#8221; model isn&#8217;t just structural. It&#8217;s conceptual. And <a href="https://sloanreview.mit.edu/projects/the-emerging-agentic-enterprise-how-leaders-must-navigate-a-new-age-of-ai/">MIT Sloan&#8217;s research</a> names the tension precisely.</p><p>Organizations have two established mental models for things that do work. <strong>Tools</strong> are owned, controlled, and predictable. You buy them, deploy them, and they perform the same function until they depreciate. <strong>Workers</strong> are autonomous, managed through contracts and oversight, capable of judgment and adaptation. You hire them, develop them, and hold them accountable.</p><p>AI agents are both simultaneously - and neither fully.</p><p>Like tools, agents are owned. There&#8217;s no employment contract, no rights negotiation, no career aspiration. They can be duplicated, shut down, or replaced with a better version overnight. Like workers, agents act autonomously. They make decisions, adapt to context, and produce outputs that weren&#8217;t explicitly programmed. They can pursue sub-goals, interact with external systems, and take actions that surprise their operators.</p><p>MIT&#8217;s framing of this tension is direct: &#8220;How do you supervise something designed to work autonomously?&#8221; The ownership model of tools says you control it. The management model of workers says you guide it. Agents require both - programmatic constraints AND ongoing oversight - which is exactly what neither HR frameworks nor IT asset management were built to provide.</p><p>The org chart metaphor collapses this into one model: treat agents like employees, give them roles and managers and performance reviews. The result is that organizations apply the wrong governance framework to entities that need a purpose-built one.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://marcosnewsletter.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">If you&#8217;re finding this useful, consider subscribing. I publish data-driven breakdowns on enterprise AI maturity every Tuesday - no hype, no vendor pitches. </p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><h2>What happens when you manage agents like employees</h2><p>The consequences of the wrong mental model are already showing up in the data. And they&#8217;re not theoretical.</p><p><a href="https://labs.cloudsecurityalliance.org/research/csa-research-note-ai-agent-governance-framework-gap-20260403/">The Cloud Security Alliance&#8217;s April 2026 research</a> on enterprise agent governance found that <strong>80% of organizations deploying AI agents have documented &#8220;risky agent behaviors&#8221;</strong> - including unauthorized system access, data exposure, and actions that exceeded intended scope. Among companies with more than $1 billion in revenue, <strong>64% reported losses exceeding $1 million from AI system failures</strong> in 2025.</p><p>The pattern is specific and instructive. <strong>34% of these incidents occurred in systems where the agent had behavioral instructions in its prompt - but zero programmatic enforcement.</strong> The agent was told what to do in natural language. Nobody built the guardrails to ensure it actually did it.</p><p>This is the employee model in action. When you onboard a human employee, you give them instructions, context, and trust. You expect them to exercise judgment within reasonable boundaries. If they make a mistake, you course-correct through feedback. The system works because human employees have social incentives, career stakes, and built-in caution about overstepping authority.</p><p>Agents have none of these natural constraints. An agent with broad database permissions will use those permissions without hesitation, without social awareness, without the instinct that tells a human &#8220;maybe I should check before running a write query on the production database.&#8221; When <a href="https://www.bcg.com/publications/2025/what-happens-ai-stops-asking-permission">BCG documented a vulnerability</a> where an AI agent had SQL write privileges to a service account exposing 3.17 trillion rows, the root cause wasn&#8217;t a technology failure. It was an organizational assumption: the agent was provisioned with the same kind of trust you&#8217;d extend to a senior employee. No human employee would ever have had unaudited write access to 3.17 trillion rows. But the employee mental model made it feel natural to grant it.</p><p>The multi-agent coordination problem is even more revealing. <a href="https://cogentinfo.com/resources/when-ai-agents-collide-multi-agent-orchestration-failure-playbook-for-2026/">Research on multi-agent deployment failures</a> found that <strong>36.9% of failures occur at coordination seams</strong> - the handoff points between agents. The failure modes are distinctly non-human: infinite loops where agents with conflicting instructions bounce tasks indefinitely, &#8220;hallucinated consensus&#8221; where agents appear to agree on fabricated data, and rigid escalation chains where outdated assumptions propagate across the system under stress.</p><p>No amount of &#8220;performance reviews&#8221; or &#8220;management feedback&#8221; addresses these failure modes. They require engineering solutions: programmatic guardrails, monitoring infrastructure, kill switches, and governance layers that operate at machine speed.</p><div><hr></div><h2>The right questions to ask about agents</h2><p>The questions that matter aren&#8217;t about managing agents like people. They&#8217;re about governing agents like systems.</p><p>Instead of &#8220;How do we give feedback to an agent?&#8221;, ask: <strong>&#8220;Have we systematized our own thinking enough to specify what &#8216;good output&#8217; means for this task?&#8221;</strong> Giving feedback to an agent is prompt engineering. The real challenge is that most organizations can&#8217;t clearly articulate what good output looks like for human employees doing the same work. If you can&#8217;t define it for a person, you definitely can&#8217;t define it for a system.</p><p>Instead of &#8220;Do agents grow like employees?&#8221;, ask: <strong>&#8220;Who has the authority to decide what an agent can do autonomously and what requires human approval?&#8221;</strong> Agents don&#8217;t grow. They get replaced by better versions. The organizational question isn&#8217;t about their development path - it&#8217;s about decision rights, escalation protocols, and the boundary between autonomous action and human oversight.</p><p>Instead of &#8220;How do we measure agent performance?&#8221;, ask: <strong>&#8220;What output metrics define success, and how quickly can a human verify them?&#8221;</strong> Agent performance isn&#8217;t measured in annual reviews. It&#8217;s measured in real-time output validation, error rates, and alignment with defined objectives. This is operational monitoring, not talent management.</p><p>These aren&#8217;t semantic distinctions. They determine what infrastructure you build. The employee model leads to HR-style governance: periodic reviews, subjective assessments, trust-based permissions. The systems model leads to operational governance: real-time monitoring, programmatic constraints, least-privilege access, automated validation. The data is clear on which approach works: organizations applying traditional governance to agentic systems <a href="https://www.thinking.inc/en/blue-ocean/agentic/enterprise-agent-governance/">miss 60-70% of agent-specific risk vectors</a> - including action authorization, decision chain accountability, and emergent multi-agent behavior.</p><div><hr></div><h2>What agent-ready organizational design actually looks like</h2><p>If agents don&#8217;t sit in org charts, where do they go?</p><p>The answer requires a shift from static hierarchy to dynamic configuration. A mature agent-ready organization doesn&#8217;t have a fixed org chart with agent positions. It has an operational architecture where teams assemble and reconfigure based on the task at hand. Today, a project team might be 2 humans and 5 agents. Tomorrow, a different project runs with 1 human and 12 agents under a different orchestration pattern. The composition is fluid because agents, unlike employees, can be instantiated and decommissioned in minutes.</p><p>This isn&#8217;t hypothetical. It&#8217;s the direction major frameworks are pointing. McKinsey&#8217;s six shifts for the agentic organization describe the structural move from &#8220;hierarchies to outcome-oriented models&#8221; where humans focus on orchestration, not task execution. New roles are emerging to match: Agent Orchestrators who manage workflows and coordinate multi-agent systems, Hybrid Managers who supervise mixed human-agent teams, and Domain Leaders who define objectives and handle edge cases. <a href="https://sloanreview.mit.edu/article/agentic-ai-at-scale-redefining-management-for-a-superhuman-workforce/">MIT Sloan&#8217;s research</a> found that 45% of organizations with extensive agentic AI adoption expect reductions in middle management layers - not because agents replaced managers, but because the coordination work that justified those layers is now handled differently.</p><p>The governance layer is equally different. <a href="https://www.gartner.com/en/newsroom/press-releases/2025-06-25-gartner-predicts-over-40-percent-of-agentic-ai-projects-will-be-canceled-by-end-of-2027">Gartner predicts</a> that by 2028, 40% of CIOs will deploy &#8220;Guardian Agents&#8221; - AI systems that monitor, challenge, and contain the actions of other agents. This isn&#8217;t agent-as-employee with a robot manager. It&#8217;s a programmatic oversight layer that inspects prompts, tool calls, and outputs in real time, enforcing policy boundaries before actions execute. Early adoption is already underway: <a href="https://www.opsinsecurity.com/blog/gartner-market-guide-guardian-agents">17% of CIOs have deployed guardian agents and 42% plan to within one year</a>.</p><p>The infrastructure that makes this work is what I call the <strong>operational scaffold</strong> - the five-layer architecture that sits between &#8220;we have an AI model&#8221; and &#8220;we have an operational AI system.&#8221; It includes input constraints (what data the agent can access), decision rules (what autonomy thresholds apply), output standardization (how results are validated), human-in-the-loop checkpoints (when a human must approve), and operational monitoring (how drift, errors, and anomalies are caught). Without this scaffold, you have an agent on the org chart. With it, you have an agent in production.</p><div><hr></div><h2>The honest counterpoint</h2><p>The employee metaphor isn&#8217;t useless. Microsoft&#8217;s <a href="https://www.microsoft.com/en-us/worklab/work-trend-index/2025-the-year-the-frontier-firm-is-born">Frontier Firm framework</a> deliberately uses &#8220;digital colleague&#8221; as a framing, and there&#8217;s a pragmatic reason: it helps organizations culturally accept a new category of workforce participant. When you tell a VP of Operations that agents are &#8220;like really good colleagues who never sleep,&#8221; adoption barriers drop. The metaphor is a communication bridge.</p><p>The danger is when you build governance on that bridge. When the CFO asks &#8220;how many FTEs can we replace with agents?&#8221;, the employee metaphor becomes a costing model that ignores inference economics, integration overhead, and the governance infrastructure that agents require and humans don&#8217;t. <a href="https://agentmodeai.com/the-hidden-costs-of-agentic-ai-a-cfos-guide-to-true-tco-and-roi-modeling/">Research on hidden AI costs</a> found that 73% of enterprises don&#8217;t understand the true total cost of ownership of AI systems, with hidden costs comprising 70% of actual investment beyond licensing.</p><p>Use the metaphor to build cultural readiness. Don&#8217;t use it to architect governance. Those are different jobs, and confusing them is how 40% of agentic AI projects end up canceled.</p><div><hr></div><h2>What to do Monday morning</h2><p>If your organization is deploying or planning AI agents, three things you can do this week.</p><p><strong>First, audit your agents&#8217; permissions.</strong> Apply least-privilege by default, not trust. If an agent has write access to a production database, ask why. If the answer is &#8220;it might need it,&#8221; that&#8217;s the employee model talking. Systems get exactly the permissions they need, documented and reviewed.</p><p><strong>Second, for every agent in production, define the output specification in writing.</strong> Not &#8220;help with customer queries.&#8221; Instead: &#8220;Respond to billing inquiries using data from the CRM, escalate refund requests over $500 to a human, never access payment card data.&#8221; If you can&#8217;t write this specification, the agent isn&#8217;t ready for production - because you haven&#8217;t clarified your own thinking about what the task requires.</p><p><strong>Third, stop calling agents &#8220;team members&#8221; in internal documentation.</strong> Language shapes thinking, and thinking shapes governance. When agents are &#8220;team members,&#8221; they get trust-based permissions. When they&#8217;re operational systems, they get engineering-grade guardrails. The second framing saves money, reduces risk, and actually works.</p><p>The org chart with robot icons is a comforting vision. It suggests that the agentic future looks like today, just with better colleagues. The research says otherwise. The future organization is structurally different - more fluid, more dynamic, more dependent on operational scaffolds than reporting lines. Building toward that future starts with letting go of the metaphor that&#8217;s holding back the design.</p><div><hr></div><p><em>Next week: the governance vacuum. NIST, the EU AI Act, and ISO 42001 have zero specific guidance on autonomous AI agents. Meanwhile, every enterprise is deploying them. What this regulatory gap means for your risk profile - and what to do before the regulators catch up.</em></p><div><hr></div><h2>Sources &amp; Further Reading</h2><ol><li><p><a href="https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai">McKinsey: The State of AI in 2025 - Agents, Innovation, and Transformation</a> (2025)</p></li><li><p><a href="https://sloanreview.mit.edu/projects/the-emerging-agentic-enterprise-how-leaders-must-navigate-a-new-age-of-ai/">MIT Sloan: The Emerging Agentic Enterprise</a> (2025-2026)</p></li><li><p><a href="https://sloanreview.mit.edu/article/agentic-ai-at-scale-redefining-management-for-a-superhuman-workforce/">MIT Sloan: Agentic AI at Scale - Redefining Management for a Superhuman Workforce</a> (2026)</p></li><li><p><a href="https://labs.cloudsecurityalliance.org/research/csa-research-note-ai-agent-governance-framework-gap-20260403/">Cloud Security Alliance: AI Agent Governance Framework Gap</a> (April 2026)</p></li><li><p><a href="https://www.bcg.com/publications/2025/what-happens-ai-stops-asking-permission">BCG: What Happens When AI Stops Asking Permission</a> (2025)</p></li><li><p><a href="https://www.gartner.com/en/newsroom/press-releases/2025-06-25-gartner-predicts-over-40-percent-of-agentic-ai-projects-will-be-canceled-by-end-of-2027">Gartner: Over 40% of Agentic AI Projects Will Be Canceled by End of 2027</a> (June 2025)</p></li><li><p><a href="https://cogentinfo.com/resources/when-ai-agents-collide-multi-agent-orchestration-failure-playbook-for-2026/">Cogent Info: When AI Agents Collide - Multi-Agent Orchestration Failure Playbook</a> (2026)</p></li><li><p><a href="https://www.thinking.inc/en/blue-ocean/agentic/enterprise-agent-governance/">Enterprise Agent Governance: Bridging the Gap</a> (2026)</p></li><li><p><a href="https://www.opsinsecurity.com/blog/gartner-market-guide-guardian-agents">Gartner Market Guide: Guardian Agents</a> (2025-2026)</p></li><li><p><a href="https://www.microsoft.com/en-us/worklab/work-trend-index/2025-the-year-the-frontier-firm-is-born">Microsoft: The Frontier Firm - Work Trend Index 2025</a> (2025)</p></li><li><p><a href="https://agentmodeai.com/the-hidden-costs-of-agentic-ai-a-cfos-guide-to-true-tco-and-roi-modeling/">Hidden Costs of Agentic AI: A CFO&#8217;s Guide to True TCO</a> (2026)</p></li></ol>]]></content:encoded></item><item><title><![CDATA[The Pilot Graveyard: $30 Billion Spent on AI Pilots. 95% Delivered Nothing.]]></title><description><![CDATA[Last week I analyzed 17 AI maturity frameworks and found none of them work. This week: the pattern that kills AI pilots before they ever reach production - and the structure that breaks the cycle.]]></description><link>https://marcosnewsletter.substack.com/p/pilot-graveyard-ai-failure-rate</link><guid isPermaLink="false">https://marcosnewsletter.substack.com/p/pilot-graveyard-ai-failure-rate</guid><dc:creator><![CDATA[Marco]]></dc:creator><pubDate>Tue, 07 Apr 2026 12:28:47 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!WE3E!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff79af426-2d08-4536-b585-e31e3268ef6f_1672x940.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Last week I made the case that <a href="https://open.substack.com/pub/marcosnewsletter/p/ai-maturity-frameworks-comparison?r=e6ccc&amp;utm_campaign=post&amp;utm_medium=web&amp;showWelcomeOnShare=true">17 AI maturity frameworks exist and none of them agree on what &#8220;mature&#8221; means</a>. The natural follow-up question is: if we can&#8217;t even measure maturity, what happens to the companies that try to build AI anyway?</p><p>The answer, according to MIT&#8217;s July 2025 <a href="https://mlq.ai/media/quarterly_decks/v0.1_State_of_AI_in_Business_2025_Report.pdf">State of AI in Business report</a>, is that <strong>95% of enterprise GenAI pilots deliver zero measurable P&amp;L impact</strong>. Not &#8220;underperform.&#8221; Not &#8220;take longer than expected.&#8221; Zero.</p><p>That&#8217;s $30-40 billion in enterprise AI spending producing nothing that shows up on the income statement. And the trend isn&#8217;t improving.</p><div><hr></div><h2>The numbers are getting worse, not better</h2><p><a href="https://www.spglobal.com/market-intelligence/en/news-insights/research/ai-experiences-rapid-adoption-but-with-mixed-outcomes-highlights-from-vote-ai-machine-learning">S&amp;P Global&#8217;s Voice of the Enterprise survey</a> found that <strong>42% of companies abandoned the majority of their AI initiatives in 2025</strong> - up from 17% the year before. That&#8217;s a 2.5x increase in one year. The average company now scraps 46% of its AI proofs-of-concept before they reach production.</p><p>Meanwhile, <a href="https://media-publications.bcg.com/The-Widening-AI-Value-Gap-Sept-2025.pdf">BCG&#8217;s September 2025 research</a> shows the gap between winners and everyone else is accelerating. The 5% of companies BCG classifies as &#8220;future-built&#8221; achieve <strong>1.7x revenue growth and 3.6x three-year total shareholder returns</strong> compared to their peers. The competitive gap has widened consistently since 2016 and shows no signs of closing.</p><p><a href="https://www.forrester.com/blogs/predictions-2026-ai-moves-from-hype-to-hard-hat-work/">Forrester predicts</a> that enterprises will defer <strong>25% of planned AI spending to 2027</strong>, with CFOs demanding stronger ROI evidence before approving budgets. The hype cycle is correcting, and the correction is painful.</p><p>The pattern is consistent across every major research house: adoption is near-universal (88% of companies use AI somewhere, per <a href="https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai">McKinsey</a>), but only 39% report any measurable business impact. The same McKinsey data shows that companies who redesign workflows around AI - rather than bolting AI onto existing processes - are 2.8x more likely to capture that value. AI isn&#8217;t failing because companies aren&#8217;t trying. It&#8217;s failing because the 95% and the 5% are trying in fundamentally different ways.</p><div><hr></div><h2>MIT found the root cause. It&#8217;s not what vendors tell you.</h2><p>The most important line in the MIT report is this: <strong>&#8220;Most GenAI systems do not retain feedback, adapt to context, or improve over time.&#8221;</strong></p><p>Read that again. The systems don&#8217;t learn. They don&#8217;t adapt. They don&#8217;t get better. Each pilot exists in isolation, disconnected from the workflow it was supposed to improve, with no feedback loop between what the AI produces and what the business actually needs.</p><p>MIT&#8217;s diagnosis isn&#8217;t about infrastructure, regulation, or talent. It&#8217;s about <strong>organizational learning</strong> - or, more precisely, the absence of it. The study found that &#8220;most GenAI pilots are disconnected from real workflows and lack meaningful feedback loops.&#8221; The pilot runs in a sandbox. Someone demonstrates it in a meeting. Everyone nods. Then it sits there, untouched, because nobody built the bridge between &#8220;technically works&#8221; and &#8220;operationally useful.&#8221;</p><p>This is the critical distinction. The failure isn&#8217;t at the model layer. The failure is at the integration layer - the point where AI capability meets organizational process. And that failure has a specific, predictable pattern:</p><p><strong>Step 1:</strong> An executive sees a demo or reads a report. Excitement builds. A team gets budget to &#8220;explore AI.&#8221;</p><p><strong>Step 2:</strong> The team picks a use case, builds a pilot, and demonstrates it. The demo looks impressive. The numbers look promising in the controlled environment.</p><p><strong>Step 3:</strong> Someone asks &#8220;how do we put this in production?&#8221; The room goes quiet. The pilot was built on sample data, not production data. The workflow it&#8217;s supposed to improve has twelve edge cases nobody mapped. The people who actually do the work weren&#8217;t involved in the design.</p><p><strong>Step 4:</strong> The pilot lingers. Updates slow down. The team gets reassigned. Six months later, a new executive arrives with a new AI initiative. The cycle restarts.</p><p>This isn&#8217;t one company&#8217;s story. This is the default enterprise experience with AI. And it compounds.</p><p>Now contrast that with what BCG&#8217;s top 5% do. Same Step 1 - an executive gets excited. But at Step 2, the team picks a use case attached to a real production workflow, not an abstract opportunity. At Step 3, production readiness isn&#8217;t an afterthought because the pilot was built against real data and real process constraints from the start. At Step 4, there&#8217;s an owner who stays accountable and a feedback loop that improves the system every week. Same starting point. Radically different sequence.</p><div><hr></div><h2>Innovation paralysis: the psychological trap nobody names</h2><p>There&#8217;s a layer beneath the operational failure that the research doesn&#8217;t fully capture, but that anyone working in enterprise AI recognizes immediately.</p><p>The AI market generates a constant stream of new tools, new models, new startups, and new claims. Every week there&#8217;s a product launch that promises to &#8220;manage your entire business&#8221; or &#8220;replace your workforce.&#8221; Every conference keynote shows a demo that looks like the future.</p><p>For enterprise decision-makers, this creates a specific form of paralysis. You can&#8217;t evaluate every new tool. You can&#8217;t pilot every promising model. You can&#8217;t keep up with the release cadence even if you tried. So you do what feels responsible: you run lots of small experiments. You &#8220;explore&#8221; broadly. You keep your options open.</p><p>The result is what I call <strong>innovation paralysis</strong> - the paradox where chasing every innovation blocks all innovation. Companies run dozens of proofs-of-concept, spend heavily on simulations and experiments, but never implement anything at scale. There&#8217;s always a newer model, a better tool, a more promising approach just around the corner.</p><p>MIT&#8217;s finding that <strong>&#8220;the normalization of perpetual piloting has become the most visible failure of 2025&#8221;</strong> names the symptom. Innovation paralysis names the cause.</p><p>The irony is that beneath all this noise, there&#8217;s a clear enterprise roadmap unfolding. Products like Google&#8217;s Gemini Enterprise, Salesforce&#8217;s Agentforce, and Microsoft&#8217;s Copilot follow real business needs, not hype cycles. The enterprise-grade adoption path is slower, more practical, and more productive than the experimental one. But it&#8217;s invisible to companies stuck chasing the next breakthrough.</p><div><hr></div><h2>Scar tissue: the compounding cost of failure</h2><p>Every failed pilot leaves a mark. Not just on the budget, but on the organization.</p><p><a href="https://www.crema.us/blog/agency-scar-tissue-why-digital-transformations-fail-and-how-to-heal-the-wounds">Crema calls this &#8220;organizational scar tissue&#8221;</a> - the lingering emotional and structural damage from failed transformation initiatives. It shows up as cynicism in meetings (&#8221;we tried AI last year and it didn&#8217;t work&#8221;), as resistance to new proposals (&#8221;show me the ROI before I&#8217;ll even consider it&#8221;), and as a general organizational immune response to anything labeled &#8220;AI.&#8221;</p><p>The math is brutal. A failed pilot costs $500k-$2M directly. But the downstream damage - two to three years of organizational resistance to the next initiative, and the next one requiring two to three times longer to build trust - compounds the real cost to tens of millions over a five-year horizon.</p><p>Transformation fatigue is the compound interest of failure. After two or three rounds of &#8220;this time it&#8217;s different&#8221; that turns out not to be different, the organization&#8217;s willingness to change drops to near zero. The problem isn&#8217;t that teams don&#8217;t believe AI works. It&#8217;s that they&#8217;ve learned to expect failure from their organization&#8217;s way of implementing it.</p><p>This is the vicious cycle: bad implementation creates scar tissue, scar tissue creates resistance, resistance blocks the next initiative, and the failure repeats.</p><p>The 5% that scale AI successfully have the opposite dynamic. Each win - even a small one - creates organizational proof that &#8220;this works for us.&#8221; That proof reduces resistance for the next initiative, which ships faster, which builds more proof. Scar tissue compounds failure. Momentum compounds success. The question is which cycle your organization is feeding.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://marcosnewsletter.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">If you&#8217;re finding this useful, consider subscribing. I publish data-driven breakdowns on enterprise AI maturity every Tuesday - no hype, no vendor pitches.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><h2>The pattern that works: narrow, fast, real</h2><p>The research converges on one structural insight: <strong>the companies that capture value from AI start narrow and scale gradually.</strong> Not narrow as in &#8220;small ambition.&#8221; Narrow as in &#8220;one specific, high-value workflow, done properly, with measurable results.&#8221;</p><p>That 2.8x multiplier I mentioned earlier - the gap between companies redesigning workflows and those just adding AI on top - comes down to three specific characteristics. The difference isn&#8217;t sophistication. It&#8217;s focus.</p><p>The pattern that works has three characteristics:</p><p><strong>1. It targets a specific pain, not a broad opportunity.</strong> &#8220;We&#8217;ll use AI to improve customer experience&#8221; fails. &#8220;We&#8217;ll reduce our average handle time in the support queue from 10 minutes to 8 minutes&#8221; succeeds. The first is a direction. The second is a measurable outcome with a clear workflow attached.</p><p>The numbers from focused implementations are consistent: customer service handle time drops by 20% ($120k/year in savings for mid-sized teams). Lead scoring conversion improves by 22% (3.2-month payback). Churn prediction accuracy improves enough to reduce churn by 9% in a single quarter (under 2-month payback).</p><p><strong>2. It connects to real workflows from day one.</strong> The biggest difference between a pilot that scales and a pilot that dies is whether it was built inside the production workflow or next to it. MIT&#8217;s core finding - pilots disconnected from real workflows don&#8217;t learn and don&#8217;t improve - has a practical implication: the pilot has to touch real data, real users, and real processes from the start.</p><p>This doesn&#8217;t mean shipping untested AI into production. It means the pilot is designed with the production environment in mind: real data inputs, real user feedback, real process constraints. A pilot that works on sample data in a sandbox but hasn&#8217;t been tested against the twelve edge cases in the actual workflow is not a pilot. It&#8217;s a demo.</p><p><strong>3. It&#8217;s designed to scale, not to prove a point.</strong> There&#8217;s a fundamental design difference between a proof-of-concept and what some practitioners call a <strong>Minimum Viable Transformation</strong> (MVT). A PoC answers the question &#8220;can this work?&#8221; An MVT answers &#8220;can this work at scale, in production, within our organizational constraints?&#8221;</p><p>The MVT includes things a typical PoC skips: pre-defined success metrics with Go/No-Go decision gates. A feedback loop that routes user corrections back into the system. A plan for what happens when the model is wrong (because it will be wrong). An owner who stays accountable past the demo.</p><p>One signal from the MIT study reinforces this: <strong>vendor-built tools succeed 67% of the time, while internal custom builds succeed only 33%.</strong> This isn&#8217;t an argument for outsourcing everything. It&#8217;s an argument for speed and specialization over customization. Focused, well-scoped implementations with existing tools outperform ambitious custom builds by a factor of two.</p><div><hr></div><h2>Designing the exit from the pilot graveyard</h2><p>The sequence matters. Organizations stuck in the pilot graveyard typically try to fix the problem by doing more of what caused it: more planning, more strategy, more comprehensive AI initiatives. The research suggests the opposite.</p><p><a href="https://www.kotterinc.com/methodology/8-steps/">John Kotter&#8217;s change management research</a> offers a useful framework here. Step 6 of his 8-step process is &#8220;Generate Short-Term Wins,&#8221; and his reasoning is precise: <strong>&#8220;Wins are the molecules of results. They must be recognized, collected, and communicated - early and often - to track progress and energize volunteers to persist.&#8221;</strong></p><p>Applied to AI: the first priority isn&#8217;t a comprehensive strategy. It&#8217;s one visible, measurable win that heals scar tissue and builds organizational confidence that &#8220;this time, it actually is different.&#8221;</p><p>The practical sequence:</p><p><strong>Week 1-2:</strong> Identify 2-3 workflows where the pain is high, the technical complexity is low, and the ROI is clearly measurable. This doesn&#8217;t require a six-month assessment. It requires someone who understands the operations and can spot where time and money are being lost. In most organizations, the people doing the actual work can tell you in 30 minutes where the worst bottlenecks are.</p><p><strong>Day 15-90:</strong> Build and deploy the first focused implementation. Not a proof-of-concept. An actual working system integrated into the real workflow, with real feedback loops, being used by real people. Measure the before and after. Publish the results internally.</p><p><strong>Day 90+:</strong> Use the momentum from the first win to scope the second. Each successful implementation reduces organizational resistance and creates advocates who pull the next project forward. This is where the scaling begins - not from a top-down mandate, but from demonstrated value.</p><p>The pattern is: Discovery (find the opportunity), Integration (embed AI in the specific workflow), Amplification (extend to adjacent processes), Transformation (rearchitect at scale). Each phase funds and justifies the next.</p><div><hr></div><h2>The honest counterpoint</h2><p>The objection to this approach is predictable: &#8220;Quick wins create silos. You need a holistic strategy and solid data foundation first.&#8221; This is a reasonable concern, and in some contexts it&#8217;s valid.</p><p>But <a href="https://exec-ed.berkeley.edu/2025/09/beyond-roi-are-we-using-the-wrong-metric-in-measuring-ai-success/">UC Berkeley&#8217;s research on AI success metrics</a> suggests the reality is more nuanced. Comprehensive strategies that never ship are the defining characteristic of pilot graveyards. McKinsey&#8217;s data shows that organizations starting with specific, high-value workflows and scaling gradually actually achieve more holistic results than those attempting comprehensive transformation from day one.</p><p>The key is intent. A quick win built as a terminal point - &#8220;we automated this one thing, done&#8221; - does create silos. A quick win built as the first step of a deliberate scaling path - with shared infrastructure, common data standards, and documented patterns - creates a foundation. The difference is in the design, not the scope.</p><p>There&#8217;s also the agent dimension. Gartner predicts that <a href="https://www.gartner.com/en/articles/ai-maturity-model">40% of agentic AI projects will be canceled by 2027</a> due to escalating costs and unclear business value. This next wave of AI complexity makes the argument for focused, validated implementations even stronger, not weaker. If companies can&#8217;t scale basic GenAI pilots, the organizational challenge of autonomous agents will be orders of magnitude harder.</p><div><hr></div><h2>What to do Monday morning</h2><p>If your organization has more than three AI pilots running right now, pick the one closest to a measurable business outcome. Audit it against three questions: Is it connected to a real production workflow? Does it have a feedback loop? Is there someone accountable for its results past the demo? If the answer to any of these is no, you&#8217;ve identified why it&#8217;s not scaling.</p><p>Then find the one workflow in your operation where the pain is highest and the solution is simplest. Start there. Measure before and after. Publish the results internally. Let the momentum do what strategy decks never could.</p><p>The pilot graveyard isn&#8217;t destiny. It&#8217;s a design flaw. And design flaws can be fixed.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!WE3E!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff79af426-2d08-4536-b585-e31e3268ef6f_1672x940.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!WE3E!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff79af426-2d08-4536-b585-e31e3268ef6f_1672x940.png 424w, https://substackcdn.com/image/fetch/$s_!WE3E!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff79af426-2d08-4536-b585-e31e3268ef6f_1672x940.png 848w, https://substackcdn.com/image/fetch/$s_!WE3E!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff79af426-2d08-4536-b585-e31e3268ef6f_1672x940.png 1272w, https://substackcdn.com/image/fetch/$s_!WE3E!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff79af426-2d08-4536-b585-e31e3268ef6f_1672x940.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!WE3E!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff79af426-2d08-4536-b585-e31e3268ef6f_1672x940.png" width="1456" height="819" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f79af426-2d08-4536-b585-e31e3268ef6f_1672x940.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:819,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2459309,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://marcosnewsletter.substack.com/i/192209107?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff79af426-2d08-4536-b585-e31e3268ef6f_1672x940.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!WE3E!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff79af426-2d08-4536-b585-e31e3268ef6f_1672x940.png 424w, https://substackcdn.com/image/fetch/$s_!WE3E!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff79af426-2d08-4536-b585-e31e3268ef6f_1672x940.png 848w, https://substackcdn.com/image/fetch/$s_!WE3E!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff79af426-2d08-4536-b585-e31e3268ef6f_1672x940.png 1272w, https://substackcdn.com/image/fetch/$s_!WE3E!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff79af426-2d08-4536-b585-e31e3268ef6f_1672x940.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><p><em>Next week: AI agents aren&#8217;t chatbots. They don&#8217;t sit in org charts. And they can&#8217;t be managed like employees. Why the &#8220;agentic org chart&#8221; going viral on LinkedIn is wrong - and what organizational design for AI agents actually looks like.</em></p><div><hr></div><h2>Sources &amp; Further Reading</h2><ol><li><p><a href="https://mlq.ai/media/quarterly_decks/v0.1_State_of_AI_in_Business_2025_Report.pdf">MIT Project NANDA: The GenAI Divide - State of AI in Business 2025</a> (July 2025)</p></li><li><p><a href="https://www.spglobal.com/market-intelligence/en/news-insights/research/ai-experiences-rapid-adoption-but-with-mixed-outcomes-highlights-from-vote-ai-machine-learning">S&amp;P Global Market Intelligence: Voice of the Enterprise - AI &amp; Machine Learning 2025</a> (March 2025)</p></li><li><p><a href="https://media-publications.bcg.com/The-Widening-AI-Value-Gap-Sept-2025.pdf">BCG: The Widening AI Value Gap</a> (September 2025)</p></li><li><p><a href="https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai">McKinsey: The State of AI in 2025</a> (2025)</p></li><li><p><a href="https://www.crema.us/blog/agency-scar-tissue-why-digital-transformations-fail-and-how-to-heal-the-wounds">Crema: Agency Scar Tissue - How To Heal From Failed Digital Transformations</a></p></li><li><p><a href="https://www.kotterinc.com/methodology/8-steps/">Dr. John Kotter: The 8-Step Process for Leading Change</a></p></li><li><p><a href="https://www.forrester.com/blogs/predictions-2026-ai-moves-from-hype-to-hard-hat-work/">Forrester: Predictions 2026 - AI Moves From Hype To Hard Hat Work</a> (October 2025)</p></li><li><p><a href="https://exec-ed.berkeley.edu/2025/09/beyond-roi-are-we-using-the-wrong-metric-in-measuring-ai-success/">UC Berkeley: Beyond ROI - Are We Using the Wrong Metric in Measuring AI Success?</a> (September 2025)</p></li></ol>]]></content:encoded></item><item><title><![CDATA[17 AI Maturity Frameworks Exist. None of Them Work. Here's Why.]]></title><description><![CDATA[I analyzed 17 AI maturity frameworks from Gartner, McKinsey, Accenture, Microsoft, and others. They don't agree on how many levels exist, what to measure, and score it. None of them covers AI agents.]]></description><link>https://marcosnewsletter.substack.com/p/ai-maturity-frameworks-comparison</link><guid isPermaLink="false">https://marcosnewsletter.substack.com/p/ai-maturity-frameworks-comparison</guid><dc:creator><![CDATA[Marco]]></dc:creator><pubDate>Tue, 31 Mar 2026 12:04:57 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!NHjQ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb78e1ef5-a53e-47ec-8704-2cd047007b51_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Last week I made the case that <a href="https://claude.ai/local_sessions/link-to-article-04">organizational maturity - not data quality - is the real reason AI projects fail</a>. The research is clear: 95% of AI pilots produce zero P&amp;L impact, and the root cause is &#8220;flawed enterprise integration,&#8221; not bad models or messy databases.</p><p>The natural next question is: how mature are we, exactly? If maturity is the problem, you need a way to measure it.</p><p>So you go looking for a framework. And you find seventeen of them.</p><p>That&#8217;s not a market. That&#8217;s a red flag.</p><div><hr></div><h2>17 frameworks, 17 definitions of &#8220;mature&#8221;</h2><p>I spent the past month pulling apart every major AI maturity framework I could find: <a href="https://www.gartner.com/en/articles/ai-maturity-model">Gartner</a>, <a href="https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai-how-organizations-are-rewiring-to-capture-value">McKinsey</a>, <a href="https://www.accenture.com/us-en/insights/artificial-intelligence/ai-maturity-and-transformation">Accenture</a>, <a href="https://learn.microsoft.com/en-us/ai/playbook/strategy/maturity-model">Microsoft</a>, <a href="https://cloud.google.com/architecture/ai-ml/ai-adoption-framework">Google Cloud</a>, <a href="https://docs.aws.amazon.com/prescriptive-guidance/latest/gen-ai-enterprise-adoption-framework/maturity-model.html">AWS</a>, <a href="https://www2.deloitte.com/us/en/insights/topics/digital-transformation/ai-and-business-transformation-study.html">Deloitte</a>, <a href="https://www.pwc.com/us/en/services/consulting/library/artificial-intelligence-predictions.html">PwC</a>, <a href="https://www.ibm.com/think/topics/ai-ladder">IBM</a>, <a href="https://cisr.mit.edu/publication/2025_0801_EnterpriseAIMaturityUpdate_WoernerSebastianWeillKaganer">MIT CISR</a>, <a href="https://www.salesforce.com/blog/agentic-maturity-model/">Salesforce</a>, <a href="https://blogs.mulesoft.com/ai/agentic-enterprise-maturity-model/">MuleSoft</a>, <a href="https://onereach.ai/resources/enterprise-ai-agent-maturity-model">OneReach</a>, <a href="https://www.bain.com/insights/building-the-foundation-for-agentic-ai/">Bain</a>, <a href="https://www.nist.gov/artificial-intelligence/ai-risk-management-framework">NIST</a>, <a href="https://www.iso.org/standard/81230.html">ISO 42001</a>, and <a href="https://genai.owasp.org/resource/ai-maturity-assessment/">OWASP</a>.</p><p>The first thing you notice is that they can&#8217;t agree on the most basic structural question: how many steps does the staircase have?</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!8b8j!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc1cc853-e646-49b0-b5a0-a93a7465cb26_650x906.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!8b8j!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc1cc853-e646-49b0-b5a0-a93a7465cb26_650x906.png 424w, https://substackcdn.com/image/fetch/$s_!8b8j!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc1cc853-e646-49b0-b5a0-a93a7465cb26_650x906.png 848w, https://substackcdn.com/image/fetch/$s_!8b8j!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc1cc853-e646-49b0-b5a0-a93a7465cb26_650x906.png 1272w, https://substackcdn.com/image/fetch/$s_!8b8j!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc1cc853-e646-49b0-b5a0-a93a7465cb26_650x906.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!8b8j!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc1cc853-e646-49b0-b5a0-a93a7465cb26_650x906.png" width="650" height="906" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/dc1cc853-e646-49b0-b5a0-a93a7465cb26_650x906.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:906,&quot;width&quot;:650,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:70764,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://marcosnewsletter.substack.com/i/192076160?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc1cc853-e646-49b0-b5a0-a93a7465cb26_650x906.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!8b8j!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc1cc853-e646-49b0-b5a0-a93a7465cb26_650x906.png 424w, https://substackcdn.com/image/fetch/$s_!8b8j!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc1cc853-e646-49b0-b5a0-a93a7465cb26_650x906.png 848w, https://substackcdn.com/image/fetch/$s_!8b8j!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc1cc853-e646-49b0-b5a0-a93a7465cb26_650x906.png 1272w, https://substackcdn.com/image/fetch/$s_!8b8j!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdc1cc853-e646-49b0-b5a0-a93a7465cb26_650x906.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Three levels, four levels, five levels, six levels. Or no levels. Or binary pass/fail. OneReach&#8217;s top level is called &#8220;Organizational AGI&#8221; - a concept that doesn&#8217;t exist yet. Gartner and Google both call their top level &#8220;Transformational&#8221; but mean completely different things by it. NIST doesn&#8217;t believe in levels at all; it uses current-state vs. target-state profiles with gap analysis. And ISO 42001 reduces the entire question to a compliance audit: you&#8217;re either certified or you&#8217;re not.</p><p><strong>If the industry can&#8217;t agree on how many steps the staircase has, you&#8217;re not climbing the same staircase.</strong></p><div><hr></div><h2>Same company, different score</h2><p>The structural disagreement goes deeper than level counts. These frameworks measure fundamentally different things.</p><p>Accenture evaluates 8 dimensions and produces a score from 0 to 100 - where the median organization lands at 36 and &#8220;AI Achievers&#8221; average 64. MIT CISR calculates a &#8220;Total AI Effectiveness&#8221; percentage. Google assigns you a phase (Tactical, Strategic, or Transformational). PwC classifies you into a stage based on responsible AI practices. ISO says you have 38 controls and you either pass the audit or you don&#8217;t.</p><p>Here&#8217;s the measurement landscape across all 17:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!gWYm!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbfdec92c-48f7-4c42-b805-2d7dfd060221_680x730.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!gWYm!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbfdec92c-48f7-4c42-b805-2d7dfd060221_680x730.png 424w, https://substackcdn.com/image/fetch/$s_!gWYm!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbfdec92c-48f7-4c42-b805-2d7dfd060221_680x730.png 848w, https://substackcdn.com/image/fetch/$s_!gWYm!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbfdec92c-48f7-4c42-b805-2d7dfd060221_680x730.png 1272w, https://substackcdn.com/image/fetch/$s_!gWYm!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbfdec92c-48f7-4c42-b805-2d7dfd060221_680x730.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!gWYm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbfdec92c-48f7-4c42-b805-2d7dfd060221_680x730.png" width="680" height="730" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bfdec92c-48f7-4c42-b805-2d7dfd060221_680x730.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:730,&quot;width&quot;:680,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:67234,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://marcosnewsletter.substack.com/i/192076160?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbfdec92c-48f7-4c42-b805-2d7dfd060221_680x730.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!gWYm!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbfdec92c-48f7-4c42-b805-2d7dfd060221_680x730.png 424w, https://substackcdn.com/image/fetch/$s_!gWYm!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbfdec92c-48f7-4c42-b805-2d7dfd060221_680x730.png 848w, https://substackcdn.com/image/fetch/$s_!gWYm!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbfdec92c-48f7-4c42-b805-2d7dfd060221_680x730.png 1272w, https://substackcdn.com/image/fetch/$s_!gWYm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbfdec92c-48f7-4c42-b805-2d7dfd060221_680x730.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Run the same organization through Accenture&#8217;s model and you might get a score of 42 - &#8220;Implemented AI.&#8221; Run it through PwC&#8217;s model and you&#8217;re &#8220;Strategic stage&#8221; because you have strong responsible AI practices. Put it through ISO 42001 and you fail certification because three of the 38 controls aren&#8217;t documented. Apply the NIST framework and there&#8217;s no score at all - just a list of gaps between where you are and where you want to be.</p><p>Four frameworks, four verdicts, zero overlap.</p><p>This matters because these assessments drive real decisions. They determine where budget goes, which projects get approved, and how leadership reports AI progress to the board. When the tool you use to measure maturity changes the measurement itself, you don&#8217;t have assessment. You have astrology with better slide decks.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://marcosnewsletter.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption"><strong>If this analysis is useful, subscribe to get the next one.</strong> Every Tuesday I publish data-driven breakdowns on enterprise AI maturity, agent strategy, and operations. No hype, no vendor pitches.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><h2>The vendor problem hiding in plain sight</h2><p>There&#8217;s a structural reason these frameworks disagree, and it&#8217;s not academic.</p><p>Salesforce&#8217;s maturity model measures readiness to adopt Agentforce. Microsoft&#8217;s agentic model measures readiness for Copilot and Azure AI Services. Google Cloud&#8217;s framework assesses progress toward Vertex AI adoption. IBM&#8217;s AI Ladder leads to watsonx. AWS&#8217;s model is explicitly about Amazon Bedrock.</p><p>These aren&#8217;t neutral assessments. They&#8217;re sales funnels with maturity language.</p><p>The pattern is consistent: score low on dimension X, and the recommended action is to buy product Y - from the same company that built the framework. <a href="https://www.accenture.com/us-en/insights/artificial-intelligence/ai-maturity-and-transformation">Accenture&#8217;s model</a>, developed with Carnegie Mellon&#8217;s SEI, is the most academically grounded of the vendor models. But even their assessment feeds directly into Accenture&#8217;s consulting and implementation services.</p><p>The independent frameworks - NIST, ISO, OWASP - don&#8217;t have this vendor bias. But they have a different problem: they&#8217;re designed for compliance and risk management, not for operational maturity assessment. NIST tells you where your risks are. ISO tells you if your management system is auditable. Neither tells you how to get from &#8220;we have a chatbot in customer service&#8221; to &#8220;AI is embedded in how we operate.&#8221;</p><p>So the vendor models measure operational readiness but point you toward their platform. The independent models avoid vendor bias but don&#8217;t measure operational readiness. Nobody covers both.</p><div><hr></div><h2>The agent-shaped hole</h2><p>All of this might be a solvable problem if the frameworks at least covered the same ground. But there&#8217;s one area where the gap is catastrophic: AI agents.</p><p><a href="https://www2.deloitte.com/us/en/insights/topics/digital-transformation/ai-and-business-transformation-study.html">Deloitte&#8217;s 2026 survey</a> of 3,235 enterprise leaders found that <strong>73% plan to deploy AI agents within two years.</strong> But only <strong>21% have governance frameworks mature enough to handle them.</strong> That&#8217;s a 52-percentage-point gap between ambition and readiness.</p><p><a href="https://www.gartner.com/en/articles/ai-agents">Gartner predicts</a> that <strong>40% of agentic AI projects will be canceled by 2027</strong> - not because the technology fails, but because organizations aren&#8217;t prepared.</p><p>And the maturity frameworks? Most were built before the agentic shift.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!fOOM!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8bffce7-5d74-44e6-9225-222dbc6da243_680x906.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!fOOM!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8bffce7-5d74-44e6-9225-222dbc6da243_680x906.png 424w, https://substackcdn.com/image/fetch/$s_!fOOM!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8bffce7-5d74-44e6-9225-222dbc6da243_680x906.png 848w, https://substackcdn.com/image/fetch/$s_!fOOM!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8bffce7-5d74-44e6-9225-222dbc6da243_680x906.png 1272w, https://substackcdn.com/image/fetch/$s_!fOOM!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8bffce7-5d74-44e6-9225-222dbc6da243_680x906.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!fOOM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8bffce7-5d74-44e6-9225-222dbc6da243_680x906.png" width="680" height="906" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f8bffce7-5d74-44e6-9225-222dbc6da243_680x906.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:906,&quot;width&quot;:680,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:76666,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://marcosnewsletter.substack.com/i/192076160?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8bffce7-5d74-44e6-9225-222dbc6da243_680x906.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!fOOM!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8bffce7-5d74-44e6-9225-222dbc6da243_680x906.png 424w, https://substackcdn.com/image/fetch/$s_!fOOM!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8bffce7-5d74-44e6-9225-222dbc6da243_680x906.png 848w, https://substackcdn.com/image/fetch/$s_!fOOM!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8bffce7-5d74-44e6-9225-222dbc6da243_680x906.png 1272w, https://substackcdn.com/image/fetch/$s_!fOOM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff8bffce7-5d74-44e6-9225-222dbc6da243_680x906.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Only Microsoft has a dedicated, structured agent maturity model. Salesforce and OneReach have agent-specific models too, but they&#8217;re tied to proprietary platforms. IBM has the most advanced agent governance tooling. OWASP covers agent security. The rest either added agent modules recently, mention agents in passing, or ignore them entirely.</p><p>Meanwhile, every enterprise is deploying agents. The frameworks were supposed to tell you if you&#8217;re ready. Most of them don&#8217;t even know you&#8217;re asking the question.</p><div><hr></div><h2>The deeper problem nobody is naming</h2><p>Even if one framework were perfect - right number of levels, right dimensions, no vendor bias, full agent coverage - it would still miss the most important signal.</p><p>Every framework produces a single, organization-wide maturity score. &#8220;You are at Level 3.&#8221; &#8220;Your AI Effectiveness is 62%.&#8221; &#8220;You are in the Strategic phase.&#8221;</p><p>That number is as misleading as GDP per capita.</p><p>An organization at &#8220;Level 2 average&#8221; where every team sits between 1.5 and 2.5 is fundamentally different from one at &#8220;Level 2 average&#8221; where the data science team is at 4, marketing is at 3, and operations and HR are at 0. The second organization has the same average score but radically different operational reality. The data science team is deploying production models. Operations is running everything manually. There&#8217;s no cross-functional synergy because half the teams can&#8217;t participate.</p><p><strong>The standard deviation between teams matters more than the mean.</strong></p><p>An organization with low average but tight variance can grow systematically. Train everyone, improve together, build processes that scale across departments. An organization with the same average but high variance has a political problem first and a maturity problem second: the advanced teams resent the laggards, the laggards feel left behind, and every cross-functional AI initiative stalls because the weakest link determines the pace.</p><p>No framework measures this. Not one of the 17 I analyzed even mentions intra-organizational variance as a dimension. They all assume the organization is a monolith that can be assigned a single score.</p><p>Ask any VP of Engineering whether every team in their org is at the same maturity level and they&#8217;ll laugh. Yet that&#8217;s exactly what these frameworks assume when they produce a single number.</p><div><hr></div><h2>What this means for you</h2><p>I&#8217;m not going to pretend there&#8217;s an easy fix here. If 17 frameworks from the world&#8217;s largest consulting firms, tech companies, and standards bodies haven&#8217;t converged on a shared definition, adding an eighteenth won&#8217;t solve it.</p><p>But there are three things worth doing right now.</p><p><strong>First, stop treating any single framework as ground truth.</strong> If you&#8217;re using Gartner&#8217;s model because &#8220;it&#8217;s Gartner,&#8221; understand that you&#8217;re getting Gartner&#8217;s definition of maturity, measured by Gartner&#8217;s dimensions, through Gartner&#8217;s lens. That&#8217;s one perspective, not the answer.</p><p><strong>Second, check the agent question.</strong> If your organization is planning to deploy AI agents - and statistically, it probably is - ask whether your current assessment framework has anything to say about it. If the answer is &#8220;not really,&#8221; your assessment is already outdated. The thing you&#8217;re about to deploy is the thing your measurement tool doesn&#8217;t cover.</p><p><strong>Third, look inside before you look outside.</strong> Before you pick a framework, ask a simpler question: <strong>do the teams in your organization even agree on what &#8220;AI mature&#8221; means?</strong> If the ML team thinks maturity is about model performance, product thinks it&#8217;s about feature adoption, and legal thinks it&#8217;s about compliance, no external framework will resolve that disagreement. It&#8217;ll just paper over it with a score.</p><p>The framework problem isn&#8217;t going to be solved by a better framework. It&#8217;s going to be solved by getting clearer about what you actually need to measure, why, and for whom.</p><p>That&#8217;s a harder conversation. It&#8217;s also a more useful one.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!NHjQ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb78e1ef5-a53e-47ec-8704-2cd047007b51_1536x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!NHjQ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb78e1ef5-a53e-47ec-8704-2cd047007b51_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!NHjQ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb78e1ef5-a53e-47ec-8704-2cd047007b51_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!NHjQ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb78e1ef5-a53e-47ec-8704-2cd047007b51_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!NHjQ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb78e1ef5-a53e-47ec-8704-2cd047007b51_1536x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!NHjQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb78e1ef5-a53e-47ec-8704-2cd047007b51_1536x1024.png" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b78e1ef5-a53e-47ec-8704-2cd047007b51_1536x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2630323,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://marcosnewsletter.substack.com/i/192076160?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb78e1ef5-a53e-47ec-8704-2cd047007b51_1536x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!NHjQ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb78e1ef5-a53e-47ec-8704-2cd047007b51_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!NHjQ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb78e1ef5-a53e-47ec-8704-2cd047007b51_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!NHjQ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb78e1ef5-a53e-47ec-8704-2cd047007b51_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!NHjQ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb78e1ef5-a53e-47ec-8704-2cd047007b51_1536x1024.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p>]]></content:encoded></item><item><title><![CDATA[The Organizational Maturity Gap Nobody Wants to Talk About]]></title><description><![CDATA[43% of companies blame data quality. 43% blame lack of technical maturity. 44% can&#8217;t name who owns AI. One of these is the symptom. The other two are the disease.]]></description><link>https://marcosnewsletter.substack.com/p/the-organizational-maturity-gap-nobody</link><guid isPermaLink="false">https://marcosnewsletter.substack.com/p/the-organizational-maturity-gap-nobody</guid><dc:creator><![CDATA[Marco]]></dc:creator><pubDate>Tue, 24 Mar 2026 12:11:44 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!ihuG!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca0a7a57-fa72-4506-9072-4b7aa3e2bd33_1424x839.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ihuG!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca0a7a57-fa72-4506-9072-4b7aa3e2bd33_1424x839.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ihuG!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca0a7a57-fa72-4506-9072-4b7aa3e2bd33_1424x839.png 424w, https://substackcdn.com/image/fetch/$s_!ihuG!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca0a7a57-fa72-4506-9072-4b7aa3e2bd33_1424x839.png 848w, https://substackcdn.com/image/fetch/$s_!ihuG!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca0a7a57-fa72-4506-9072-4b7aa3e2bd33_1424x839.png 1272w, https://substackcdn.com/image/fetch/$s_!ihuG!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca0a7a57-fa72-4506-9072-4b7aa3e2bd33_1424x839.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ihuG!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca0a7a57-fa72-4506-9072-4b7aa3e2bd33_1424x839.png" width="1424" height="839" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ca0a7a57-fa72-4506-9072-4b7aa3e2bd33_1424x839.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:839,&quot;width&quot;:1424,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:575170,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://marcosnewsletter.substack.com/i/190536710?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca0a7a57-fa72-4506-9072-4b7aa3e2bd33_1424x839.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ihuG!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca0a7a57-fa72-4506-9072-4b7aa3e2bd33_1424x839.png 424w, https://substackcdn.com/image/fetch/$s_!ihuG!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca0a7a57-fa72-4506-9072-4b7aa3e2bd33_1424x839.png 848w, https://substackcdn.com/image/fetch/$s_!ihuG!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca0a7a57-fa72-4506-9072-4b7aa3e2bd33_1424x839.png 1272w, https://substackcdn.com/image/fetch/$s_!ihuG!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fca0a7a57-fa72-4506-9072-4b7aa3e2bd33_1424x839.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p>Your data isn&#8217;t the problem.</p><p>That&#8217;s the conclusion MIT landed on after studying 350+ employees across 300+ deployments and conducting 150 interviews at companies trying to scale AI. The researchers expected to find the usual suspects: messy data, siloed systems, legacy infrastructure. Instead, they found something more fundamental. <strong>The root cause of <a href="https://fortune.com/2025/08/18/mit-report-95-percent-generative-ai-pilots-at-companies-failing-cfo">95% of AI pilots producing zero P&amp;L impact</a> was &#8220;flawed enterprise integration.&#8221;</strong> Not bad data. Not weak models. Not missing tools. The failure point was organizational.</p><p><a href="https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai-how-organizations-are-rewiring-to-capture-value">McKinsey&#8217;s latest state-of-AI research</a> confirms it. Only <strong>6% of companies</strong> are actually capturing value from their AI investments at scale. The gap between that 6% and the <a href="https://www.bcg.com/publications/2025/are-you-generating-value-from-ai-the-widening-gap">60% getting zero value</a> isn&#8217;t technology. It&#8217;s maturity. How companies redesign workflows, allocate budget, clarify ownership, and weave AI into their operating model. That&#8217;s all org design, not data science.</p><p>Yet walk into most vendor conversations and you&#8217;ll hear the same refrain: fix your data, implement our platform, hire our consulting team. Data quality is real (63% of companies lack AI-ready data). But it&#8217;s not the primary blocker. It&#8217;s a symptom of something deeper: unclear ownership, no data governance, and organizational structures that weren&#8217;t built for AI decision-making.</p><div><hr></div><h2>The evidence is there. We&#8217;re just looking at the wrong things.</h2><p><a href="https://www.pertamapartners.com/insights/ai-project-failure-statistics-2026">S&amp;P Global</a> asked enterprise leaders why their AI initiatives underperform. Here&#8217;s what they found:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!y4w5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5556b3d0-6300-4978-b7c1-71d1d49824e9_710x480.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!y4w5!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5556b3d0-6300-4978-b7c1-71d1d49824e9_710x480.png 424w, https://substackcdn.com/image/fetch/$s_!y4w5!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5556b3d0-6300-4978-b7c1-71d1d49824e9_710x480.png 848w, https://substackcdn.com/image/fetch/$s_!y4w5!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5556b3d0-6300-4978-b7c1-71d1d49824e9_710x480.png 1272w, https://substackcdn.com/image/fetch/$s_!y4w5!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5556b3d0-6300-4978-b7c1-71d1d49824e9_710x480.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!y4w5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5556b3d0-6300-4978-b7c1-71d1d49824e9_710x480.png" width="710" height="480" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5556b3d0-6300-4978-b7c1-71d1d49824e9_710x480.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:480,&quot;width&quot;:710,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:30964,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://marcosnewsletter.substack.com/i/190536710?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5556b3d0-6300-4978-b7c1-71d1d49824e9_710x480.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!y4w5!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5556b3d0-6300-4978-b7c1-71d1d49824e9_710x480.png 424w, https://substackcdn.com/image/fetch/$s_!y4w5!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5556b3d0-6300-4978-b7c1-71d1d49824e9_710x480.png 848w, https://substackcdn.com/image/fetch/$s_!y4w5!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5556b3d0-6300-4978-b7c1-71d1d49824e9_710x480.png 1272w, https://substackcdn.com/image/fetch/$s_!y4w5!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5556b3d0-6300-4978-b7c1-71d1d49824e9_710x480.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Read that table carefully. Data quality and technical maturity are cited at identical rates. And nearly half the room can&#8217;t name who owns AI in their organization. That&#8217;s not a technology problem. That&#8217;s an operating problem.</p><p>The concerning part: these percentages have gotten worse, not better. S&amp;P Global found that <strong>42% of companies have abandoned AI initiatives</strong> in the past two years, up from 17% two years prior. When you ask what they&#8217;re abandoning, it&#8217;s not projects that failed because of messy databases. It&#8217;s projects that failed because no one agreed on what success looked like, who was responsible for delivery, or how the work fit into existing operational rhythms.</p><div><hr></div><h2>MIT&#8217;s &#8220;flawed enterprise integration&#8221;: the real diagnosis</h2><p>The MIT finding deserves to sit in your head. Researchers at <a href="https://cisr.mit.edu/publication/2025_0801_EnterpriseAIMaturityUpdate_WoernerSebastianWeillKaganer">MIT&#8217;s Center for Information Systems Research</a> studied why so many AI pilots generate impressive demos but zero business impact. They tracked 300+ deployments and found that the organizations that successfully moved pilots to production had one thing in common: they <strong>redesigned how AI fit into their existing workflows, team structures, and decision-making processes.</strong></p><p>The ones that stalled had not. They treated AI as a technology insertion problem. Bolted it onto the side of existing operations. Left the same decision-makers in charge. Didn&#8217;t change how handoffs happened. Didn&#8217;t redefine what a &#8220;success metric&#8221; meant for a transformed process.</p><p>That&#8217;s &#8220;flawed enterprise integration.&#8221; The technology works. The data is fine. The problem is the organization isn&#8217;t designed to use it.</p><p>This explains why one of the starkest statistics in AI adoption research barely gets mentioned: <strong>67% of companies that bought pre-built AI solutions from vendors reported positive ROI, versus 33% that built internally.</strong> That gap isn&#8217;t because vendor solutions are technologically superior. It&#8217;s because vendors force organizational structure. They come in, audit your processes, demand you assign an owner, require you to define success upfront, and build governance into their onboarding. You have no choice but to integrate AI into your org design. Internal teams? They often just add another tool to the workflow.</p><div><hr></div><h2>High performers do five things differently</h2><p>McKinsey studied which companies actually capture value from AI and found a stark split. The top 6% aren&#8217;t smarter. They&#8217;re just organized differently.</p><p><strong>Workflow redesign:</strong> 55% of high performers redesign core workflows when deploying AI, compared to 20% of others. That&#8217;s 2.8x higher. They treat AI not as a tool to bolt on, but as a reason to rethink how work gets done. Instead of &#8220;let&#8217;s use AI to optimize what we&#8217;re already doing,&#8221; they ask &#8220;how should we do this differently?&#8221;</p><p><strong>Budget allocation:</strong> High performers allocate <strong>70% of their AI budget to people and process redesign</strong>, with 30% to technology. The rest of the industry does the opposite: 93% to technology, 7% to people.</p><p><strong>AI ambition:</strong> High performers are 3.6x more likely to pursue <strong>transformational ambition</strong> with AI rather than optimization. They&#8217;re not trying to shave 10% off a process. They&#8217;re trying to fundamentally change their competitive position.</p><p><strong>Digital budget concentration:</strong> High performers are 5x more likely to allocate 20% or more of their total digital budget to AI. They&#8217;re not dabbling. They&#8217;re committing organizational resources at a scale that forces integration.</p><p><strong>Governance clarity:</strong> Almost universally, high performers have explicitly defined who owns AI decisions, what success looks like, and how it connects to business metrics before they deploy.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!6uE3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F886a9981-0f36-43af-b547-b25dacd760a8_780x432.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!6uE3!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F886a9981-0f36-43af-b547-b25dacd760a8_780x432.png 424w, https://substackcdn.com/image/fetch/$s_!6uE3!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F886a9981-0f36-43af-b547-b25dacd760a8_780x432.png 848w, https://substackcdn.com/image/fetch/$s_!6uE3!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F886a9981-0f36-43af-b547-b25dacd760a8_780x432.png 1272w, https://substackcdn.com/image/fetch/$s_!6uE3!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F886a9981-0f36-43af-b547-b25dacd760a8_780x432.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!6uE3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F886a9981-0f36-43af-b547-b25dacd760a8_780x432.png" width="780" height="432" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/886a9981-0f36-43af-b547-b25dacd760a8_780x432.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:432,&quot;width&quot;:780,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:35474,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://marcosnewsletter.substack.com/i/190536710?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F886a9981-0f36-43af-b547-b25dacd760a8_780x432.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!6uE3!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F886a9981-0f36-43af-b547-b25dacd760a8_780x432.png 424w, https://substackcdn.com/image/fetch/$s_!6uE3!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F886a9981-0f36-43af-b547-b25dacd760a8_780x432.png 848w, https://substackcdn.com/image/fetch/$s_!6uE3!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F886a9981-0f36-43af-b547-b25dacd760a8_780x432.png 1272w, https://substackcdn.com/image/fetch/$s_!6uE3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F886a9981-0f36-43af-b547-b25dacd760a8_780x432.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><blockquote><p><strong>If this analysis is useful, subscribe to get the next one.</strong> Every Tuesday I publish data-driven breakdowns on enterprise AI maturity, agent strategy, and operations. No hype, no vendor pitches.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://marcosnewsletter.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://marcosnewsletter.substack.com/subscribe?"><span>Subscribe now</span></a></p></blockquote><div><hr></div><h2>The maturity stages: where most companies are stuck</h2><p>MIT&#8217;s Center for Information Systems Research created a four-stage maturity model for enterprise AI. Where is your organization?</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ZExH!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F105f557f-bbf6-4c75-a52e-4795d4797a9f_960x466.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ZExH!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F105f557f-bbf6-4c75-a52e-4795d4797a9f_960x466.png 424w, https://substackcdn.com/image/fetch/$s_!ZExH!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F105f557f-bbf6-4c75-a52e-4795d4797a9f_960x466.png 848w, https://substackcdn.com/image/fetch/$s_!ZExH!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F105f557f-bbf6-4c75-a52e-4795d4797a9f_960x466.png 1272w, https://substackcdn.com/image/fetch/$s_!ZExH!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F105f557f-bbf6-4c75-a52e-4795d4797a9f_960x466.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ZExH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F105f557f-bbf6-4c75-a52e-4795d4797a9f_960x466.png" width="960" height="466" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/105f557f-bbf6-4c75-a52e-4795d4797a9f_960x466.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:466,&quot;width&quot;:960,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:55743,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://marcosnewsletter.substack.com/i/190536710?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F105f557f-bbf6-4c75-a52e-4795d4797a9f_960x466.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ZExH!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F105f557f-bbf6-4c75-a52e-4795d4797a9f_960x466.png 424w, https://substackcdn.com/image/fetch/$s_!ZExH!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F105f557f-bbf6-4c75-a52e-4795d4797a9f_960x466.png 848w, https://substackcdn.com/image/fetch/$s_!ZExH!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F105f557f-bbf6-4c75-a52e-4795d4797a9f_960x466.png 1272w, https://substackcdn.com/image/fetch/$s_!ZExH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F105f557f-bbf6-4c75-a52e-4795d4797a9f_960x466.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Notice the problem: <strong>62% of companies are in Stages 1 or 2.</strong> They&#8217;re below the performance baseline. The jump from Stage 2 to Stage 3 is where financial performance starts to recover. Yet one-third of companies never make it there. They stay in the &#8220;managed pilots&#8221; phase indefinitely, spending millions on tooling and getting nothing in return.</p><p>The 7% in Stage 4 aren&#8217;t there because they have better data or smarter engineers. They got there because they rebuilt their operating model around AI. They defined clear roles. They aligned incentives. They made decisions faster. They integrated AI into how work actually gets done.</p><div><hr></div><h2>The governance multiplier nobody talks about</h2><p>Google Cloud and the <a href="https://cloudsecurityalliance.org/press-releases/2025/12/18/csa-and-google-cloud-study-finds-governance-maturity-is-strongest-predictor-of-ai-readiness">Cloud Security Alliance</a> studied the relationship between governance maturity and AI adoption outcomes. They found that companies with <strong>comprehensive governance frameworks are 2x more likely to adopt agentic AI</strong> successfully (46% adoption vs 25% for companies with basic governance). They&#8217;re also 45% more likely to keep projects running for 3+ years versus 20% for others.</p><p>The catch: only 26% of companies have comprehensive governance today. Most are flying blind. They deployed an LLM-powered chatbot somewhere, nobody documented the decision process, nobody owns the model&#8217;s performance, nobody has a way to shut it down if it breaks, and six months later it&#8217;s doing things nobody intended.</p><p><strong>Governance isn&#8217;t bureaucracy. It&#8217;s the system that lets you move fast safely.</strong> It&#8217;s what lets you deploy more, not less. High performers use governance to eliminate meetings, clarify decisions, and move projects from pilot to production in months instead of years.</p><div><hr></div><h2>The data quality trap: it&#8217;s a symptom, not the disease</h2><p>Let&#8217;s be direct about data quality because it deserves honesty.</p><p>Yes, 63% of enterprises say their data isn&#8217;t AI-ready. Yes, data quality is hard. Yes, your databases are probably messy. But here&#8217;s what&#8217;s really happening: companies with unclear ownership have no one responsible for managing data standards. Companies without governance have no mechanism to enforce data quality. Companies that haven&#8217;t redesigned workflows have no clear definition of what &#8220;ready&#8221; even means.</p><p>Fix the org design, and data quality improves.</p><p>Why? Because when someone actually owns AI outcomes (which forces them to own the data pipeline), when there&#8217;s governance that requires metadata and lineage documentation, when processes are redesigned around data rather than legacy workflows, data quality stops being a nice-to-have and becomes a required input.</p><p>The companies that fixed data quality first and expected organizational benefits got almost nothing. The ones that redesigned their org and treated data quality as part of that redesign actually succeeded. The causality flows backward from what vendors want you to believe.</p><div><hr></div><h2>The perception gap: you can&#8217;t fix what you can&#8217;t see</h2><p><a href="https://www.bcg.com/publications/2025/are-you-generating-value-from-ai-the-widening-gap">BCG found</a> that <strong>55% of organizations think they&#8217;re further along the AI maturity journey than they actually are.</strong> Put another way, they&#8217;re less advanced than they believe. And it&#8217;s not off by a little. Half the room believes they&#8217;re at Stage 3 when they&#8217;re actually in Stage 2. Some believe they&#8217;re Stage 3 when they&#8217;re barely past Stage 1.</p><p><strong>How does this happen? Easy.</strong></p><ul><li><p>When you buy an enterprise AI platform, you see yourself as an &#8220;AI adopter.&#8221;</p></li><li><p>When you run a successful pilot, you see yourself as &#8220;deploying AI.&#8221;</p></li><li><p>When a team somewhere uses an LLM in production, you see yourself as &#8220;integrated.&#8221; </p></li></ul><p><em><strong>But none of that means your organization is designed for AI</strong>. It just means you have AI running somewhere.</em></p><p>The maturity stages aren&#8217;t about whether AI exists in your company. They&#8217;re about whether your organizational structure, governance, decision-making processes, and budget allocation are designed around AI. Most companies have the first without the second.</p><div><hr></div><h2>What to do Monday morning</h2><p><strong>If you lead operations:</strong> Run a maturity audit. Not with a vendor. With your own team. Where is your organization actually in this progression? Don&#8217;t estimate. Look at real metrics: how many workflows have been redesigned? How many projects have moved from pilot to production? How many people wake up on Monday with AI ownership in their job description? Be honest. Most companies discover they&#8217;re one stage behind where they thought.</p><p><strong>If you lead engineering:</strong> Assess the vendor paradox. If you&#8217;ve been building internally and getting nowhere, and vendors consistently outperform, that&#8217;s not a technical signal. That&#8217;s an organizational one. Consider whether buying a complete solution forces the structural change you need. Sometimes the answer is &#8220;yes.&#8221;</p><p><strong>If you lead AI/ML:</strong> Map your maturity stage honestly. Then identify which transition is blocking you. Most get stuck on Stage 2 to Stage 3 (the shift from &#8220;pilots&#8221; to &#8220;integrated operations&#8221;). That transition requires redesigning workflows, not upgrading models. Figure out which workflow redesign, if successful, would unlock the most value. That&#8217;s your first move.</p><p><strong>Three diagnostic questions for everyone:</strong></p><ol><li><p>Can three different people in your organization tell you who owns AI outcomes? If not, you don&#8217;t have ownership. You have sponsors.</p></li><li><p>If a major AI initiative failed tomorrow, how many people would feel it in their quarterly results? If fewer than five, it&#8217;s not integrated enough. It&#8217;s still a side project.</p></li><li><p>What percentage of your digital budget goes to AI? Compare that to what percentage of your strategy is about AI. If there&#8217;s a gap of more than 5%, you&#8217;re not committing organizational resources proportional to the bet.</p></li></ol><div><hr></div><p>The 95% failure rate isn&#8217;t a data problem or a model problem. The 6% that succeed aren&#8217;t smarter. They&#8217;re organized differently. They&#8217;ve redesigned their operating model around AI instead of bolting AI onto an existing one.</p><p>That&#8217;s hard. It requires admitting that the org charts don&#8217;t fit anymore. It requires clarifying ownership when that&#8217;s uncomfortable. It requires redesigning workflows instead of optimizing them. It requires moving budget from technology to people.</p><p>But it&#8217;s doable. And it&#8217;s the only path that actually generates value.</p><p>Stop blaming data. Start redesigning the organization.</p><div class="native-video-embed" data-component-name="VideoPlaceholder" data-attrs="{&quot;mediaUploadId&quot;:&quot;5bd0d92d-40d9-406c-8b09-e9f74adf492b&quot;,&quot;duration&quot;:null}"></div><p></p>]]></content:encoded></item><item><title><![CDATA[40% of Agentic AI Projects Will Be Canceled. Here’s Why Yours Is Probably One of Them.]]></title><description><![CDATA[48% of enterprises plan to add $2M+ to agentic AI budgets this year. 8.6% have agents in production. 40% of projects will be canceled by 2027. The math doesn&#8217;t work - and here&#8217;s exactly where it break]]></description><link>https://marcosnewsletter.substack.com/p/40-of-agentic-ai-projects-will-be</link><guid isPermaLink="false">https://marcosnewsletter.substack.com/p/40-of-agentic-ai-projects-will-be</guid><dc:creator><![CDATA[Marco]]></dc:creator><pubDate>Tue, 17 Mar 2026 12:05:41 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!PuyW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc08738ed-8005-4cda-b520-8c6cf528db0a_2752x1536.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!PuyW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc08738ed-8005-4cda-b520-8c6cf528db0a_2752x1536.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!PuyW!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc08738ed-8005-4cda-b520-8c6cf528db0a_2752x1536.png 424w, https://substackcdn.com/image/fetch/$s_!PuyW!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc08738ed-8005-4cda-b520-8c6cf528db0a_2752x1536.png 848w, https://substackcdn.com/image/fetch/$s_!PuyW!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc08738ed-8005-4cda-b520-8c6cf528db0a_2752x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!PuyW!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc08738ed-8005-4cda-b520-8c6cf528db0a_2752x1536.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!PuyW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc08738ed-8005-4cda-b520-8c6cf528db0a_2752x1536.png" width="1456" height="813" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c08738ed-8005-4cda-b520-8c6cf528db0a_2752x1536.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:813,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:5491417,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://marcosnewsletter.substack.com/i/190152431?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc08738ed-8005-4cda-b520-8c6cf528db0a_2752x1536.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!PuyW!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc08738ed-8005-4cda-b520-8c6cf528db0a_2752x1536.png 424w, https://substackcdn.com/image/fetch/$s_!PuyW!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc08738ed-8005-4cda-b520-8c6cf528db0a_2752x1536.png 848w, https://substackcdn.com/image/fetch/$s_!PuyW!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc08738ed-8005-4cda-b520-8c6cf528db0a_2752x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!PuyW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc08738ed-8005-4cda-b520-8c6cf528db0a_2752x1536.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>48% of enterprises plan to add at least <a href="https://axis-intelligence.com/agentic-ai-adoption-statistics-2026/">$2M to agentic AI budgets</a> this year. Executive interest is at record levels. Every vendor pitch now includes the word &#8220;agent.&#8221;</p><p>The production reality is different. <a href="https://globenewswire.com/news-release/2026/03/05/3250181/31982/en/Only-7-of-Enterprises-Say-Their-Data-Is-Completely-Ready-for-AI-According-to-New-Report-from-Cloudera-and-Harvard-Business-Review-Analytic-Services.html">8.6% of companies have production-ready AI agents</a>, according to a Cloudera and Harvard Business Review Analytic Services survey of 120,000+ enterprise respondents. Not 8.6% experimenting, not 8.6% piloting. In production, processing real data, making real decisions.</p><p>Gartner predicts <a href="https://www.gartner.com/en/newsroom/press-releases/2025-06-25-gartner-predicts-over-40-percent-of-agentic-ai-projects-will-be-canceled-by-end-of-2027">40% of agentic AI projects will be canceled by 2027</a>. The three reasons are specific: escalating costs, unclear business value, and inadequate risk controls.</p><p>Those aren&#8217;t vague failure modes. They&#8217;re structural problems with names, causes, and observable patterns. Here&#8217;s what they look like up close.</p><div><hr></div><h2>Failure mode 1: The cost explosion nobody budgeted for</h2><p>Enterprise teams budget for AI agents the way they budget for SaaS subscriptions: licensing fees, infrastructure, maybe some integration work. The actual cost structure of agentic AI is fundamentally different.</p><p>An AI agent doesn&#8217;t make one call to a model. It reasons iteratively. It calls tools. It checks its own work, revises, and retries. A simple customer query that takes 100 tokens in a single-pass chatbot <a href="https://medium.com/@ajayverma23/optimizing-genai-and-agentic-ai-balancing-cost-quality-and-latency-in-production-0242ef84a306">expands to 2,000-5,000 tokens</a> when an agent processes it with tool calls, chain-of-thought reasoning, and output validation. At scale, monthly token bills dwarf infrastructure spend.</p><p>The growth is non-linear. <a href="https://online.stevens.edu/blog/hidden-economics-ai-agents-token-costs-latency/">A reasoning loop running 10 cycles consumes 50x the tokens</a> of a single linear pass. This isn&#8217;t a scaling problem you can forecast from a pilot. Pilots run on curated inputs, happy paths, and manageable volumes. Production runs on messy data, edge cases, and thousands of concurrent users. The cost curve bends sharply upward at exactly the point where you&#8217;ve already committed to the architecture.</p><p>Research from <a href="https://galileo.ai/blog/hidden-cost-of-agentic-ai">Galileo AI</a> puts the number at <strong>$5-8 per task</strong> for unconstrained agents solving software engineering issues. That&#8217;s per task, not per user per month. For a customer support operation handling 10,000 tickets per day, the economics get painful fast.</p><p><a href="https://www.eccouncil.org/cybersecurity-exchange/ai-program-manager/enterprise-ai-budget-traps/">85% of organizations misestimate AI project costs by more than 10%</a>. Nearly 25% miss forecasts by more than 50%. For agentic systems specifically, <a href="https://hypersense-software.com/blog/2026/01/12/hidden-costs-ai-agent-development/">the typical budget underestimation is 40-60%</a> when you account for the full total cost of ownership: monitoring, guardrails, human oversight, security maintenance, and ongoing tuning.</p><p>The pattern is consistent. Enterprises apply GenAI cost assumptions to agentic systems - simple LLM pricing models - when the actual cost includes orchestrators, governance layers, multiple agents running in tandem, and operational overhead that scales faster than usage. By the time the real costs surface, the project is already deep in production and the sunk-cost fallacy kicks in.</p><div><hr></div><h2>Failure mode 2: Agent-washing and the vendor trust gap</h2><p>The agentic AI market is experiencing a labeling crisis. Every automation tool, chatbot, and workflow builder now calls itself an &#8220;agent.&#8221; The term has become meaningless as a product descriptor.</p><p>Of the thousands of vendors now marketing agentic AI capabilities, <a href="https://www.outreach.io/resources/blog/agent-washing-ai-projects-fail-guide">approximately 130 are building actual agentic systems</a> - tools that can reason, adapt to novel situations, maintain state across interactions, and make autonomous decisions within defined boundaries. The rest are rebranding scripted automation, RPA flows, and basic chatbots. Analysts have started calling it &#8220;agent-washing.&#8221;</p><p>The adoption numbers reflect this confusion. <a href="https://www.deloitte.com/us/en/insights/topics/technology-management/tech-trends/2026/agentic-ai-strategy.html">30% of enterprises say they&#8217;re exploring agentic AI. 38% report they&#8217;re piloting. But only 11% have agents actively running in production</a>. The gap between &#8220;we&#8217;re piloting&#8221; and &#8220;it&#8217;s in production&#8221; is the largest in any enterprise technology category right now. For comparison, cloud adoption saw a 2-3 year gap. Agentic AI may see 4-5 years, because the infrastructure requirements are harder to meet.</p><p>The consequence for enterprise buyers is expensive. Teams purchase tools marketed as &#8220;agentic,&#8221; build workflows around them, discover the tools can&#8217;t actually handle production complexity, and either rebuild or abandon. The CIO team funded an &#8220;agentic AI initiative.&#8221; What they got was a chatbot with a nicer interface.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://marcosnewsletter.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption"><strong>If this analysis is useful, subscribe to get the next one.</strong> Every Tuesday I publish data-driven breakdowns on enterprise AI maturity, agent strategy, and operations. No hype, no vendor pitches.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><h2>Failure mode 3: The governance vacuum</h2><p>This is the failure mode that kills projects after they&#8217;re built rather than before. And it&#8217;s the one most enterprises are least prepared for.</p><p><a href="https://secureprivacy.ai/blog/ai-risk-compliance-2026/">78% of enterprises now embed AI in their strategic plans</a>. Only 19% have fully implemented governance frameworks. That&#8217;s a 4x gap between ambition and controls - and it widens for agentic systems, which require governance architectures that most frameworks haven&#8217;t even defined yet.</p><p>The numbers at the operational level are worse. <a href="https://hackernoon.com/agentic-ai-governance-frameworks-2026-risks-oversight-and-emerging-standards">64% of organizations experimenting with agentic AI</a> have no formal agent monitoring or escalation protocols. Agents are making decisions in test environments with no structure for what happens when those decisions affect real customers, real revenue, or real compliance obligations.</p><p><a href="https://globenewswire.com/news-release/2026/03/05/3250181/31982/en/Only-7-of-Enterprises-Say-Their-Data-Is-Completely-Ready-for-AI-According-to-New-Report-from-Cloudera-and-Harvard-Business-Review-Analytic-Services.html">Only 7% of enterprises report their data is completely ready for AI adoption</a>. 27% say their data is not ready at all. Agents need reliable, structured, governed data to make good decisions. When they get bad data, they don&#8217;t just produce bad output - they take bad actions. At enterprise scale, a single agent making autonomous decisions on corrupted data can cascade through downstream systems before anyone notices.</p><p>The incidents are already happening. <a href="https://www.helpnetsecurity.com/2026/03/03/enterprise-ai-agent-security-2026/">64% of companies with $1B+ annual revenue have lost more than $1M to AI-related failures</a>. These aren&#8217;t hypothetical risks from research papers. They&#8217;re production incidents where agents with insufficient guardrails took actions they shouldn&#8217;t have, accessed data they shouldn&#8217;t have, or made decisions that nobody had designed an override for.</p><p>The regulatory environment is catching up, but enterprises aren&#8217;t ready for it. The <a href="https://www.venable.com/insights/publications/2026/02/agentic-ai-is-here-legal-compliance-and-governance">EU AI Act&#8217;s high-risk compliance deadline is August 2, 2026</a>. The <a href="https://www.mayerbrown.com/en/insights/publications/2026/02/governance-of-agentic-artificial-intelligence-systems">EU Product Liability Directive</a>, with its implementation deadline of December 9, 2026, explicitly classifies software and AI systems as &#8220;products&#8221; subject to strict liability for defects. When an agent fails and causes damage, someone is legally responsible. Most enterprises haven&#8217;t determined who that is.</p><div><hr></div><h2>What the 6% do differently</h2><p>McKinsey&#8217;s State of AI data identifies a small group of &#8220;AI high performers&#8221; - organizations attributing 5% or more of EBIT to AI. They represent roughly 6% of the companies surveyed. Their practices are distinct and measurable.</p><p><strong>They start with governance, not technology.</strong> A <a href="https://cloudsecurityalliance.org/press-releases/2025/12/18/csa-and-google-cloud-study-finds-governance-maturity-is-strongest-predictor-of-ai-readiness">Cloud Security Alliance and Google Cloud study</a> found that governance maturity is the single strongest predictor of AI readiness. Organizations with comprehensive governance policies are <a href="https://www.dataversity.net/articles/building-a-practical-framework-for-ai-governance-maturity-in-the-enterprise/">2x more likely to adopt agentic AI</a> than those with partial guidelines, and nearly 4x more likely than those with developing policies. The adoption rates are striking: 46% early adoption among companies with comprehensive governance, 25% with partial guidelines, 12% with developing policies.</p><p>This reverses the common sequence. Most enterprises buy tools first and build governance later. The 6% build governance first and buy tools that fit within it.</p><p><strong>They invest in people and processes, not just software.</strong> <a href="https://blog.exceeds.ai/ai-budget-allocation-tracking/">High-performing organizations allocate roughly 70% of AI budgets to people, processes, and organizational readiness</a> - not tools. 40-50% goes specifically to talent and training in early stages. The typical enterprise inverts this ratio, spending 60-70% on software and infrastructure and hoping the organizational change follows.</p><p><strong>They pursue transformational change, not efficiency.</strong> McKinsey&#8217;s data shows high performers are <a href="https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai">3x more likely to pursue transformational change</a> rather than incremental efficiency gains. They don&#8217;t ask &#8220;how can AI make this process faster?&#8221; They ask &#8220;should this process exist in its current form?&#8221; That question leads to workflow redesign, which is the single strongest correlate with AI-driven EBIT impact.</p><p><strong>They build confidence through structure.</strong> Organizations with mature governance report <a href="https://cloudsecurityalliance.org/blog/2025/12/18/ai-security-governance-your-maturity-multiplier">48% confidence in protecting AI systems</a>, compared to 16-23% for those with partial or developing frameworks. This 2-3x confidence gap matters because it determines whether leadership approves production deployment or keeps projects in permanent pilot mode. Governance doesn&#8217;t slow you down. It&#8217;s the prerequisite for moving fast with institutional support.</p><div><hr></div><h2>The regulatory clock most enterprises ignore</h2><p>The governance vacuum isn&#8217;t just an operational risk. It&#8217;s becoming a legal one on a specific timeline.</p><p><strong>August 2, 2026</strong>: EU AI Act high-risk compliance deadline. Organizations deploying AI in areas classified as high-risk - hiring, credit scoring, law enforcement, critical infrastructure - must have conformity assessments, risk management systems, and human oversight in place.</p><p><strong>December 9, 2026</strong>: EU Product Liability Directive implementation deadline. AI systems are explicitly classified as &#8220;products.&#8221; Defects that cause damage trigger strict liability - the standard applied to manufacturing defects, not the negligence standard typically applied to software.</p><p><strong>Already active</strong>: The FTC&#8217;s <a href="https://secureprivacy.ai/blog/ai-risk-compliance-2026/">Operation AI Comply</a> is targeting deceptive AI marketing claims. If your agent doesn&#8217;t do what you tell customers it does, enforcement is already underway.</p><p>In the U.S., regulatory fragmentation persists - no unified federal framework, with the FTC, NIST, and Department of Commerce interpreting AI within existing mandates. <a href="https://www.venable.com/insights/publications/2026/02/agentic-ai-is-here-legal-compliance-and-governance">Colorado&#8217;s AI Act takes effect June 30, 2026</a>. Singapore released a <a href="https://hackernoon.com/agentic-ai-governance-frameworks-2026-risks-oversight-and-emerging-standards">draft governance framework for agentic AI</a> at Davos in January.</p><p>The direction is clear even where the specifics vary: regulation is converging on accountability for autonomous AI systems. Enterprises that haven&#8217;t built governance architectures by mid-2026 will be retrofitting under deadline pressure - the most expensive and error-prone way to comply.</p><div><hr></div><h2>The honest counterpoint</h2><p>The failure narrative is real, but so are the acceleration signals.</p><p>Production agent deployment <a href="https://codewave.com/insights/ai-enterprise-adoption-2026/">nearly doubled in four months</a> - from 7.2% to 13.2% between August and December 2025. Infrastructure costs are falling: NVIDIA&#8217;s Blackwell architecture enables significant inference cost reductions, and open-source models are driving token costs down. <a href="https://www.pwc.com/us/en/tech-effect/ai-analytics/ai-agent-survey.html">66% of organizations adopting agents report measurable productivity value</a>, according to PwC.</p><p>These signals matter, but they need context. &#8220;Measurable productivity value&#8221; is a lower bar than P&amp;L impact. Inference cost reductions don&#8217;t account for the full operational cost of production agents - orchestration, monitoring, guardrails, and human oversight. And the deployment acceleration, while encouraging, is starting from a base of 7.2%. Doubling to 13.2% still leaves 87% of enterprises without production agents.</p><p>The optimistic case is that the infrastructure is maturing and costs are declining. The realistic case is that infrastructure maturity and cost reduction are necessary conditions, not sufficient ones. Without governance, organizational readiness, and honest assessment of total costs, cheaper tools just mean cheaper failures.</p><div><hr></div><h2>Diagnosing your position</h2><p>Five questions that separate the 8.6% from the 91.4%:</p><p><strong>1. How many of your AI agents are in production, processing real data, making real decisions?</strong> Not in a sandbox, not with curated inputs, not monitored by the team that built them. In production, with production data, under normal business operations. If the answer is zero, your agentic AI initiative is still an experiment regardless of how much you&#8217;ve invested.</p><p><strong>2. Can you quantify your actual per-task cost for any deployed agent?</strong> Not the LLM API pricing from the vendor&#8217;s website. The actual cost: tokens consumed per task including reasoning loops, tool calls, retries, and failed attempts. Infrastructure. Monitoring. Human review time. If you don&#8217;t know this number, you can&#8217;t forecast whether the system is economically viable at scale.</p><p><strong>3. Do you have a governance framework that exists independently of any specific AI tool or vendor?</strong> Validation rules, escalation policies, audit trails, human-in-the-loop patterns. If your governance is embedded in a vendor&#8217;s platform, you don&#8217;t have governance - you have a feature that disappears when you switch vendors.</p><p><strong>4. Who is accountable when an agent makes a wrong decision in production?</strong> Not &#8220;who built the agent&#8221; or &#8220;who approved the project.&#8221; Who is legally and operationally accountable for the downstream impact? If the answer takes more than five seconds, you have a liability gap.</p><p><strong>5. What&#8217;s your timeline for EU AI Act compliance?</strong> If you&#8217;re deploying agents in high-risk categories and don&#8217;t have a conformity assessment plan, you&#8217;re five months from a regulatory deadline with no roadmap.</p><div class="native-video-embed" data-component-name="VideoPlaceholder" data-attrs="{&quot;mediaUploadId&quot;:&quot;76e93c4a-b09b-4f88-b819-288c8005baa9&quot;,&quot;duration&quot;:null}"></div><p></p><div><hr></div><h2>What to do Monday morning</h2><p><strong>If you lead operations:</strong> Pull the actual cost data on your agent deployments. Not the budget, not the forecast - the actual spend. Compare per-task costs to the value each task delivers. If you can&#8217;t do this calculation, that&#8217;s your first problem. Build the measurement before you scale the system.</p><p><strong>If you lead engineering:</strong> Audit your vendor stack for agent-washing. For each tool marketed as &#8220;agentic,&#8221; test whether it can handle a novel situation it wasn&#8217;t explicitly designed for. Can it reason about an edge case? Can it recover from an unexpected input? If the answer is no, you have an automation tool, not an agent. Plan accordingly.</p><p><strong>If you lead AI governance:</strong> Map the gap between your current governance framework and the EU AI Act&#8217;s high-risk requirements. Then assess whether your framework covers agentic-specific risks: autonomous decision-making, multi-agent coordination, cascading failures, data access scope. If it doesn&#8217;t, you have five months to close that gap.</p><p>The pattern in the data is unambiguous. Enterprises that build governance before deploying agents, invest in organizational readiness before buying tools, and rigorously assess costs before scaling are the ones landing in the 6%. Everyone else is funding the 40% cancellation rate.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!jQOl!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa810e8ec-8a5d-4efc-ad2a-2cf52559eab7_1536x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!jQOl!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa810e8ec-8a5d-4efc-ad2a-2cf52559eab7_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!jQOl!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa810e8ec-8a5d-4efc-ad2a-2cf52559eab7_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!jQOl!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa810e8ec-8a5d-4efc-ad2a-2cf52559eab7_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!jQOl!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa810e8ec-8a5d-4efc-ad2a-2cf52559eab7_1536x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!jQOl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa810e8ec-8a5d-4efc-ad2a-2cf52559eab7_1536x1024.png" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a810e8ec-8a5d-4efc-ad2a-2cf52559eab7_1536x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:719852,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://marcosnewsletter.substack.com/i/190152431?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa810e8ec-8a5d-4efc-ad2a-2cf52559eab7_1536x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!jQOl!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa810e8ec-8a5d-4efc-ad2a-2cf52559eab7_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!jQOl!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa810e8ec-8a5d-4efc-ad2a-2cf52559eab7_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!jQOl!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa810e8ec-8a5d-4efc-ad2a-2cf52559eab7_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!jQOl!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa810e8ec-8a5d-4efc-ad2a-2cf52559eab7_1536x1024.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p>]]></content:encoded></item><item><title><![CDATA[Your AI Roadmap Is a Single Line. It Should Be Three - (The Three Horizons of AI Adoption)]]></title><description><![CDATA[Most enterprises plan AI as a single track: experiment, then scale. The companies capturing value run three horizons in parallel.]]></description><link>https://marcosnewsletter.substack.com/p/your-ai-roadmap-is-a-single-line</link><guid isPermaLink="false">https://marcosnewsletter.substack.com/p/your-ai-roadmap-is-a-single-line</guid><dc:creator><![CDATA[Marco]]></dc:creator><pubDate>Mon, 09 Mar 2026 06:02:41 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!byG3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38c971c1-6e87-4894-9a51-dea40e090d19_911x518.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!byG3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38c971c1-6e87-4894-9a51-dea40e090d19_911x518.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!byG3!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38c971c1-6e87-4894-9a51-dea40e090d19_911x518.png 424w, https://substackcdn.com/image/fetch/$s_!byG3!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38c971c1-6e87-4894-9a51-dea40e090d19_911x518.png 848w, https://substackcdn.com/image/fetch/$s_!byG3!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38c971c1-6e87-4894-9a51-dea40e090d19_911x518.png 1272w, https://substackcdn.com/image/fetch/$s_!byG3!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38c971c1-6e87-4894-9a51-dea40e090d19_911x518.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!byG3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38c971c1-6e87-4894-9a51-dea40e090d19_911x518.png" width="911" height="518" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/38c971c1-6e87-4894-9a51-dea40e090d19_911x518.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:518,&quot;width&quot;:911,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:285752,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://marcosnewsletter.substack.com/i/190126609?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38c971c1-6e87-4894-9a51-dea40e090d19_911x518.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!byG3!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38c971c1-6e87-4894-9a51-dea40e090d19_911x518.png 424w, https://substackcdn.com/image/fetch/$s_!byG3!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38c971c1-6e87-4894-9a51-dea40e090d19_911x518.png 848w, https://substackcdn.com/image/fetch/$s_!byG3!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38c971c1-6e87-4894-9a51-dea40e090d19_911x518.png 1272w, https://substackcdn.com/image/fetch/$s_!byG3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F38c971c1-6e87-4894-9a51-dea40e090d19_911x518.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In a <a href="https://open.substack.com/pub/marcosnewsletter/p/from-chatgpt-to-operational-ai-the?r=e6ccc&amp;utm_campaign=post&amp;utm_medium=web&amp;showWelcomeOnShare=true">previous post</a>, I described the gap between passive AI (employees using ChatGPT) and operational AI (AI embedded in business processes with scaffolding). The scaffold - input constraints, decision rules, output standardization, human-in-the-loop, monitoring - is what separates a working demo from a production system.</p><p>That post answered the question &#8220;what do I need to build?&#8221; This one answers &#8220;when do I build it - and what comes after?&#8221;</p><p>The short answer: you need three parallel tracks, not one sequential plan. Most enterprises are running a single-horizon AI strategy. The 6% capturing real value are running three.</p><div class="native-video-embed" data-component-name="VideoPlaceholder" data-attrs="{&quot;mediaUploadId&quot;:&quot;34c7565f-bfa3-4559-a51c-937551c420fc&quot;,&quot;duration&quot;:null}"></div><div><hr></div><h2>The 88-6 gap</h2><p><a href="https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai">McKinsey&#8217;s 2025 State of AI</a> reports that 88% of organizations use AI in at least one function. Generative AI adoption specifically jumped from 33% to 72% in a single year. By any adoption metric, enterprise AI is mainstream.</p><p>Value capture is not. Only 6% of organizations qualify as &#8220;AI high performers&#8221; - those attributing 5% or more of EBIT to AI. The other 82% have adopted the technology without capturing the economics.</p><p>The usual explanation is execution: companies aren&#8217;t implementing well enough. But the <a href="https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai">McKinsey data</a> points to something more structural. Among high performers, the strongest differentiator isn&#8217;t model sophistication or talent density. It&#8217;s <strong>workflow redesign</strong> - 55% of high performers fundamentally redesign business processes around AI, compared to 20% of everyone else. High performers are also 5x more likely to allocate 20%+ of their digital budget to AI.</p><p>These aren&#8217;t companies that experiment better. They plan differently. They treat AI not as a tool to layer onto existing processes but as infrastructure that requires new processes, new roles, and new governance - built across multiple time horizons simultaneously.</p><div><hr></div><h2>Three horizons, defined</h2><p>McKinsey&#8217;s <a href="https://www.itonics-innovation.com/blog/702010-rule-of-innovation">Three Horizons framework</a> was originally designed for innovation portfolio management. It maps well to enterprise AI because the fundamental challenge is the same: how to allocate resources across current operations, near-term transformation, and long-term disruption.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://marcosnewsletter.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Marco's Substack is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>Applied to AI adoption:</p><p><strong>Horizon 1: Experimentation.</strong> Employees use AI tools for individual productivity. ChatGPT, Copilot, Gemini. Pilots run in sandboxes with curated data. Value is real but localized - faster emails, quicker code reviews, better first drafts. No integration with core systems. No governance beyond acceptable use policies.</p><p><strong>Horizon 2: Operational AI.</strong> AI is embedded in business processes. It receives structured inputs, applies business rules, produces standardized outputs, and connects to downstream systems. Humans validate at checkpoints, not at every step. This is the scaffold layer - the architecture that turns ad hoc usage into repeatable, governed workflows.</p><p><strong>Horizon 3: Agentic Systems at Scale.</strong> Multiple AI agents coordinate across domains. An orchestrator delegates tasks to specialist agents. Agents monitor other agents (guardian agents). The system handles cross-functional workflows autonomously, with human oversight at strategic decision points rather than operational ones.</p><p>H1: Experimentation H2: Operational AI H3: Agentic Systems <strong>AI role</strong> Utility (human decides) System component (AI acts within constraints) Autonomous coordinator (AI orchestrates) <strong>Integration</strong> None - browser tab Embedded in specific workflows Cross-system, cross-domain <strong>Data requirements</strong> Whatever the user pastes in Structured pipelines, validated inputs Unified data fabric, real-time access <strong>Governance</strong> Acceptable use policy Business rules, HITL validation, audit trails Guardian agents, policy enforcement, observability <strong>People</strong> Individual users Process owners, AI operators Agent orchestrators, AI governance specialists <strong>Typical timeline</strong> 0-6 months 6-18 months 18-36 months</p><p>The horizons aren&#8217;t sequential stages where you finish one before starting the next. They&#8217;re parallel investment tracks. The infrastructure you build in H1 and H2 becomes the foundation for H3. Skip or delay H2 investment, and H3 becomes impossible to execute on timeline.</p><div><hr></div><h2>Where everyone actually is</h2><p>The distribution is heavily skewed toward H1. According to <a href="https://www.gartner.com/en/newsroom/press-releases/2025-08-26-gartner-predicts-40-percent-of-enterprise-apps-will-feature-task-specific-ai-agents-by-2026-up-from-less-than-5-percent-in-2025">Gartner&#8217;s 2025 AI adoption data</a>, 62% of organizations are experimenting with AI agents, 23% are scaling them in at least one function, and <a href="https://www.pwc.com/us/en/industries/tmt/library/trust-and-safety-outlook/rise-and-risks-of-agentic-ai.html">only 11% have agentic AI in production</a>.</p><p>Most enterprise &#8220;AI roadmaps&#8221; are really H1 extension plans: more use cases for ChatGPT, more pilots in more departments, more prompt engineering workshops. The roadmap moves laterally across the organization rather than vertically through the horizons. Companies add breadth (more experiments) without adding depth (operational infrastructure).</p><p><a href="https://fortune.com/2025/08/18/mit-report-95-percent-generative-ai-pilots-at-companies-failing-cfo/">MIT&#8217;s 2025 research</a> puts numbers on the cost of this approach: 95% of AI pilots delivered no measurable P&amp;L impact. 88% never reached production. The gap between a working pilot and an operational system is precisely the H1-to-H2 transition - and most organizations have no plan for crossing it.</p><p>The failure modes are predictable. Data quality issues surface when pilots meet real systems (43% of failures). Integration complexity is underestimated by 2-3x (31% of technical failures). Use cases are chosen for technical novelty rather than business value (43% of strategic failures). These aren&#8217;t pilot problems. They&#8217;re infrastructure problems that only become visible when you try to move from H1 to H2.</p><div><hr></div><h2>The infrastructure dependency chain</h2><p>Here&#8217;s why the horizons can&#8217;t be sequential: what you build in H2 is a prerequisite for H3, and what you build in H1 should be designed with H2 in mind.</p><p><a href="https://mitsloan.mit.edu/ideas-made-to-matter/5-heavy-lifts-deploying-ai-agents">MIT Sloan research</a> found that 80% of the effort in deploying AI agents is consumed by data engineering, stakeholder alignment, and governance - not model development. The model is the easy part. The infrastructure is the hard part. And infrastructure takes time to build, test, and harden.</p><p>Consider what H3 (agentic systems) requires that can&#8217;t be retrofitted:</p><p><strong>Governed data access.</strong> Agents need real-time access to data across systems. If your H1 pilots run on manually curated datasets and your production data lives in silos, there&#8217;s no shortcut. <a href="https://www.informatica.com/resources/articles/enterprise-ai-agent-engineering.html">90% of organizations with AI workflows in production</a> use an integration platform. The other 10% built their own. Either way, it takes 12-18 months to do well.</p><p><strong>Governance frameworks.</strong> An employee using ChatGPT who gets a bad answer wastes five minutes. An autonomous agent with bad data executes dozens of flawed decisions before anyone notices. The governance infrastructure - validation rules, escalation policies, audit trails, human-in-the-loop patterns - must exist before agents get decision authority. You build this in H2, not H3.</p><p><strong>Observability and monitoring.</strong> <a href="https://www.informatica.com/resources/articles/enterprise-ai-agent-engineering.html">91% of AI models experience quality degradation over time</a>. For H3 systems where agents act autonomously, degradation detection must be automated. <a href="https://www.gartner.com/en/newsroom/press-releases/2025-06-11-gartner-predicts-that-guardian-agents-will-capture-10-15-percent-of-the-agentic-ai-market-by-2030">Gartner predicts</a> that guardian agents - AI systems that monitor, challenge, and contain other agents - will capture 10-15% of the agentic AI market by 2030. But the monitoring patterns start in H2, with simpler operational systems.</p><p><strong>API-led architecture.</strong> Traditional enterprise systems were designed for deterministic, synchronous workflows. Agents need asynchronous communication, state management across boundaries, and conflict resolution. Event-driven architectures and API-first design are H2 investments that become H3 requirements.</p><p>Each of these takes time. Companies that start building them only when they&#8217;re &#8220;ready for H3&#8221; are already 12-24 months behind.</p><div><hr></div><h2>The 70-20-10 allocation</h2><p>McKinsey&#8217;s original Three Horizons framework suggests a resource allocation of roughly 70% to H1 (core operations), 20% to H2 (adjacent growth), and 10% to H3 (transformational bets). Applied to AI investment, the ratios shift - but the principle holds.</p><p>Most enterprises are running something closer to 95-5-0: nearly all resources on experimentation and pilots, minimal investment in operational infrastructure, nothing on Agentic architecture.</p><p>High performers, based on the <a href="https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai">McKinsey data</a>, allocate differently. They spend 20%+ of digital budgets on AI (5x the median), and they direct that spending toward workflow redesign and integration - not more pilots. A reasonable allocation for a company serious about all three horizons might look like:</p><p><strong>H1 (50-60%): Keep experimenting, but with structure.</strong> Run pilots that are designed to test H2 infrastructure, not just model capabilities. Every pilot should answer: &#8220;What would it take to put this in production?&#8221; If the answer is &#8220;rebuild everything,&#8221; the pilot is testing the wrong thing.</p><p><strong>H2 (25-35%): Build the scaffold.</strong> Data pipelines, governance frameworks, HITL patterns, API connectivity, monitoring. This is the unsexy work that makes everything else possible. Fund it like infrastructure, not like innovation.</p><p><strong>H3 (10-15%): Start small, start now.</strong> Prototype multi-agent patterns. Evaluate orchestration frameworks. Build a guardian agent for one high-risk process. You&#8217;re not deploying at scale - you&#8217;re building institutional knowledge about what H3 will require.</p><p>The key insight from the high-performer data isn&#8217;t that they spend more on AI. It&#8217;s that they spend differently - investing in the operational layers that connect experimentation to business impact.</p><div><hr></div><h2>2026 is the inflection year</h2><p>Multiple signals converge on 2026 as a watershed for enterprise AI.</p><p><a href="https://www.gartner.com/en/newsroom/press-releases/2025-08-26-gartner-predicts-40-percent-of-enterprise-apps-will-feature-task-specific-ai-agents-by-2026-up-from-less-than-5-percent-in-2025">Gartner predicts</a> that 40% of enterprise applications will feature task-specific AI agents by 2026, up from less than 5% in 2025. The <a href="https://www.deloitte.com/us/en/insights/industry/technology/technology-media-and-telecom-predictions/2026/ai-agent-orchestration.html">autonomous AI agent market</a> could reach $8.5 billion this year. And <a href="https://www.gartner.com/en/newsroom/press-releases/2025-06-25-gartner-predicts-over-40-percent-of-agentic-ai-projects-will-be-canceled-by-end-of-2027">Gartner separately predicts</a> that 40% of agentic AI projects will be canceled by end of 2027 due to escalating costs, unclear business value, or inadequate risk controls.</p><p>Read those cancellation drivers: escalating costs (no infrastructure planning), unclear value (no workflow redesign), inadequate risk controls (no governance). They&#8217;re symptoms of companies jumping to H3 without building H2.</p><p>The deadline isn&#8217;t technical - it&#8217;s financial. CFOs have funded two years of experimentation. <a href="https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai">92% of companies</a> plan to increase AI investment over the next three years. But the patience for indefinite pilots is ending. Companies that haven&#8217;t moved from experimentation to operations by late 2026 will face pressure to justify continued AI spending against zero P&amp;L impact.</p><p>A caveat worth noting: much of what gets labeled &#8220;Agentic&#8221; in 2026 will be supervised AI operating within tight constraints - closer to advanced H2 than true H3 autonomy. <a href="https://www.gartner.com/en/newsroom/press-releases/2025-06-25-gartner-predicts-over-40-percent-of-agentic-ai-projects-will-be-canceled-by-end-of-2027">Gartner estimates</a> that only about 130 of thousands of &#8220;Agentic AI vendors&#8221; offer real Agentic capabilities. The rest is rebranding - what analysts are calling &#8220;agent washing.&#8221; This doesn&#8217;t change the planning framework. If anything, it reinforces it: the companies that built real operational infrastructure will be positioned to adopt real agentic capabilities. The ones chasing vendor demos will discover they bought H1 tools dressed up as H3.</p><div><hr></div><h2>What to do Monday morning</h2><p>Three diagnostic questions:</p><ol><li><p><strong>What percentage of your AI usage is in production, processing real data, integrated into real workflows?</strong> If less than 15%, your roadmap is an H1 plan regardless of what it says.</p></li><li><p><strong>Do you have a governance framework - validation rules, escalation policies, audit trails - that exists independently of any specific AI tool?</strong> If not, you have no H2 foundation. Every Agentic project you start will need to build this from scratch.</p></li><li><p><strong>Who owns the H2-to-H3 transition?</strong> Not &#8220;who sponsors AI&#8221; or &#8220;who runs pilots.&#8221; Who is responsible for the operational infrastructure that connects experimentation to autonomous systems? If that person doesn&#8217;t exist, the transition won&#8217;t happen.</p></li></ol><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://marcosnewsletter.substack.com/p/your-ai-roadmap-is-a-single-line/comments&quot;,&quot;text&quot;:&quot;Leave a comment&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://marcosnewsletter.substack.com/p/your-ai-roadmap-is-a-single-line/comments"><span>Leave a comment</span></a></p><p>Three actions:</p><p><strong>If you lead operations:</strong> Audit your AI pilots for production-readiness. For each pilot, document what&#8217;s missing: data pipeline, governance, integration, monitoring. That gap list is your H2 investment plan.</p><p><strong>If you lead engineering:</strong> Design your next integration project with agent compatibility in mind. Event-driven architecture, API-first design, bounded scopes with clear permissions. These are H2 investments that compound in H3.</p><p><strong>If you lead ML/AI:</strong> Stop optimizing models in isolation. Partner with ops and engineering on the scaffold - the validation rules, the HITL patterns, the monitoring infrastructure. <a href="https://mitsloan.mit.edu/ideas-made-to-matter/5-heavy-lifts-deploying-ai-agents">80% of the work that makes agents operational</a> is the part your team probably isn&#8217;t doing.</p><p>The AI adoption race is over. Most enterprises have given their people access to the tools. The next race - who builds the operational infrastructure that makes those tools deliver business value - is the one that determines who&#8217;s in the 6% and who&#8217;s in the 82% that adopted AI and got nothing for it.</p><p></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://marcosnewsletter.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Marco's Substack is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[From ChatGPT to Operational AI: The Architecture Shift Nobody Plans For]]></title><description><![CDATA[Most enterprises have mastered AI adoption. Almost none have built the infrastructure to make it operational.]]></description><link>https://marcosnewsletter.substack.com/p/from-chatgpt-to-operational-ai-the</link><guid isPermaLink="false">https://marcosnewsletter.substack.com/p/from-chatgpt-to-operational-ai-the</guid><dc:creator><![CDATA[Marco]]></dc:creator><pubDate>Fri, 06 Mar 2026 16:29:53 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!nf5D!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F070f4e1c-746a-432c-ba0d-272efe4a025f_1059x575.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!nf5D!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F070f4e1c-746a-432c-ba0d-272efe4a025f_1059x575.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!nf5D!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F070f4e1c-746a-432c-ba0d-272efe4a025f_1059x575.png 424w, https://substackcdn.com/image/fetch/$s_!nf5D!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F070f4e1c-746a-432c-ba0d-272efe4a025f_1059x575.png 848w, https://substackcdn.com/image/fetch/$s_!nf5D!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F070f4e1c-746a-432c-ba0d-272efe4a025f_1059x575.png 1272w, https://substackcdn.com/image/fetch/$s_!nf5D!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F070f4e1c-746a-432c-ba0d-272efe4a025f_1059x575.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!nf5D!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F070f4e1c-746a-432c-ba0d-272efe4a025f_1059x575.png" width="1059" height="575" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/070f4e1c-746a-432c-ba0d-272efe4a025f_1059x575.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:575,&quot;width&quot;:1059,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:415366,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://marcosnewsletter.substack.com/i/190109433?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7d9c3e81-9ca9-4e9d-8027-10c3d3f954a7_1222x765.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!nf5D!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F070f4e1c-746a-432c-ba0d-272efe4a025f_1059x575.png 424w, https://substackcdn.com/image/fetch/$s_!nf5D!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F070f4e1c-746a-432c-ba0d-272efe4a025f_1059x575.png 848w, https://substackcdn.com/image/fetch/$s_!nf5D!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F070f4e1c-746a-432c-ba0d-272efe4a025f_1059x575.png 1272w, https://substackcdn.com/image/fetch/$s_!nf5D!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F070f4e1c-746a-432c-ba0d-272efe4a025f_1059x575.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>ChatGPT Enterprise usage grew 8x last year. The tool reached <a href="https://openai.com/index/the-state-of-enterprise-ai-2025-report/">92% Fortune 500 penetration</a> in under 24 months. Employees are writing emails, summarizing documents, generating code. Adoption, by any measure, is working.</p><p>Business impact is not. Only <a href="https://www.forrester.com/blogs/predictions-2026-the-shift-from-ai-hype-to-hard-business-outcomes/">13% of companies report positive EBITDA impact</a> from AI. Fewer than one-third of enterprise decision-makers can tie AI investments to P&amp;L growth. And 42% of companies scrapped most of their AI initiatives in 2025 - up from 17% the year before.</p><p>The gap between adoption and impact has a structural explanation. Organizations have built <strong>passive AI</strong> - employees using language models ad hoc - and assumed that business results would follow. They don&#8217;t. What follows is a plateau: high usage, shallow integration, no measurable return.</p><p>The missing piece isn&#8217;t a better model. It&#8217;s an <strong>operational scaffold</strong> - the architecture that turns ad hoc AI usage into repeatable, governed, integrated business processes. Most teams never build it. The ones that do scale 3x faster than their peers.</p><div class="native-video-embed" data-component-name="VideoPlaceholder" data-attrs="{&quot;mediaUploadId&quot;:&quot;0c21e7da-bcc5-4c65-b331-91b8edbe9942&quot;,&quot;duration&quot;:null}"></div><p></p><div><hr></div><h2>Passive AI and Operational AI are different systems</h2><p>The enterprise AI conversation collapses two fundamentally different things into one.</p><p><strong>Passive AI</strong> is a utility. An employee opens ChatGPT, asks a question, gets an answer, decides what to do with it. The human remains the decision-maker. The AI is a tool, like a calculator or a search engine. If the output is wrong, the human catches it (or doesn&#8217;t - the stakes are usually low).</p><p><strong>Operational AI</strong> is a system. AI is integrated into a business process. It receives structured inputs, applies decision rules, produces standardized outputs, and feeds results into downstream workflows. Humans validate at checkpoints, not at every step. The AI has some degree of authority to act.</p><p>Passive AI Operational AI <strong>Who decides</strong> Human AI (within constraints) <strong>Integration</strong> None - browser tab Embedded in workflows and systems <strong>Input</strong> Free-form prompts Structured, validated data <strong>Output</strong> Unstructured text Standardized, schema-validated <strong>Failure mode</strong> Human catches bad output System catches bad output (or doesn&#8217;t) <strong>Infrastructure needed</strong> Login credentials Scaffold: rules, validation, HITL, monitoring</p><p>This distinction matters because the infrastructure requirements are completely different. Passive AI needs almost nothing - give people access and they&#8217;ll use it. Operational AI needs an entire architecture layer that most organizations haven&#8217;t designed, let alone built.</p><p>The <a href="https://openai.com/index/the-state-of-enterprise-ai-2025-report/">OpenAI Enterprise report</a> shows this gap in practice: 40% of knowledge workers use AI tools personally, but describe them as unreliable in enterprise systems. Among ChatGPT Enterprise users, 19% have never used data analysis features, 14% have never tried reasoning tools. Usage is wide but shallow - passive by default.</p><div><hr></div><h2>The pilot-to-production cliff</h2><p>The data on AI pilot failures is brutal.</p><p><a href="https://fortune.com/2025/08/18/mit-report-95-percent-generative-ai-pilots-at-companies-failing-cfo/">95% of enterprise gen-AI pilots deliver zero measurable return</a> on P&amp;L, according to MIT&#8217;s 2025 State of AI in Business report. 88% of pilots never reach production. The average organization abandoned 46% of its proof-of-concepts before they became operational.</p><p>These aren&#8217;t model failures. Pilots typically work fine in controlled environments with clean data and motivated teams. The failure happens at the transition - when a working prototype meets real systems, real data quality issues, real compliance requirements, and real edge cases.</p><p>The root causes, <a href="https://dotkonnekt.com/blogs/enterprise-ai/why-ai-pilots-fail-strategic-insights-to-drive-enterprise-ai-adoption">as documented across multiple enterprise studies</a>, are structural:</p><p><strong>Data quality and governance (43% of failures).</strong> Pilots succeed on manually curated datasets. Production requires integration with live systems where data is inconsistent, incomplete, and owned by nobody in particular.</p><p><strong>Integration complexity (31% of technical failures).</strong> Pilots run in isolation. Production requires API connectivity, workflow integration, compliance alignment, and authentication across multiple systems. Teams consistently underestimate this by 2-3x.</p><p><strong>Strategic misalignment (43% of failures).</strong> Use cases chosen for technical novelty rather than business value. No clear ROI defined before funding, no clear owner after launch.</p><p><strong>Organizational readiness (66% cite this as a barrier).</strong> Lack of AI skills and leadership clarity. But this is necessary, not sufficient - skilled teams fail too when the infrastructure isn&#8217;t there.</p><p>The pattern is consistent: pilots work because humans compensate for missing infrastructure. Production fails because at scale, humans can&#8217;t compensate fast enough.</p><div><hr></div><h2>What an operational scaffold actually looks like</h2><p>An operational scaffold is the infrastructure layer between &#8220;the model works&#8221; and &#8220;the system runs in production.&#8221; It&#8217;s not a single tool or framework. It&#8217;s five interconnected layers that constrain, validate, and govern AI behavior within a business process.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://marcosnewsletter.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Marco's Substack is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p><strong>Layer 1: Input constraints.</strong> What data can the AI access? Is the input valid, complete, in the right format? Are there access controls that enforce data boundaries? This layer prevents garbage-in problems and limits the AI&#8217;s exposure to data it shouldn&#8217;t process.</p><p><strong>Layer 2: Decision rules and guardrails.</strong> What business rules must the AI follow? Which actions can it take autonomously, and which require escalation? At what confidence threshold does it hand off to a human? This is where domain knowledge gets encoded - not in the prompt, but in the system architecture.</p><p><strong>Layer 3: Output standardization.</strong> Does the output match the expected schema? Does it pass quality checks? What happens when validation fails - retry, fallback to a simpler approach, or escalate? This layer makes AI output consumable by downstream systems and humans.</p><p><strong>Layer 4: Human-in-the-loop validation.</strong> Where are the checkpoints? What approval workflows exist for high-stakes decisions? How do humans provide feedback that improves the system over time? What audit trail exists for compliance and debugging?</p><p><strong>Layer 5: Operational infrastructure.</strong> How do you monitor the system in production? How do you keep data pipelines fresh? How do you version the system and roll back when something breaks? This is the layer that keeps operational AI operational - not just on launch day, but on day 200.</p><p><a href="https://www.mckinsey.com/featured-insights/mckinsey-explainers/what-are-ai-guardrails/">McKinsey&#8217;s AI guardrails framework</a> maps a similar architecture: input guardrails, runtime policy enforcement, output validation, and configuration controls. The terminology varies across frameworks. The structure is consistent.</p><p>None of these layers are optional. Skip input validation and your model processes corrupted data. Skip decision rules and the AI takes actions outside its authority. Skip output standardization and downstream systems can&#8217;t consume the results. Skip HITL and you lose the ability to catch systematic errors. Skip monitoring and you won&#8217;t know when the system degrades.</p><p>Most teams build the model. Some build one or two of these layers. Almost nobody builds all five before launching a pilot. And then they wonder why the pilot doesn&#8217;t scale.</p><div><hr></div><h2>How I built this for monthly business reviews</h2><p>I&#8217;ll make this concrete with a system I built: an operational AI pipeline for Monthly Business Reviews (MBRs) in a professional services organization.</p><p>The old process: a team spent roughly 20 hours per month collecting data from Salesforce and project management tools, validating numbers across regions, writing narrative descriptions for each KPI, and formatting everything into a leadership-ready document. The work was manual, error-prone, and repetitive.</p><p>The operational AI system reduced this to roughly 2 hours - a 90% reduction. But the model was maybe 15% of the work. The other 85% was the scaffold.</p><p><strong>Input layer.</strong> Structured data extraction from source systems with validation rules - 10 specific checks that flag inconsistencies before any narrative generation starts. Regional sums must match totals. Month-over-month trends must be directionally consistent with the underlying data. Missing data gets flagged, not silently ignored.</p><p><strong>Decision rules.</strong> The system applies business-specific logic: what counts as an improvement vs. a decline, which KPIs are bonus-linked and require extra scrutiny, how to weight regional performance against global targets. These rules aren&#8217;t in the prompt - they&#8217;re encoded in the validation pipeline.</p><p><strong>Output standardization.</strong> Every narrative follows a defined structure: current performance, trend context, root cause analysis, forward-looking implications. The format is consistent across KPIs and months, making it consumable by leadership without retraining their reading patterns.</p><p><strong>Human-in-the-loop.</strong> A data validation report goes to the operator before narratives generate. The operator reviews flagged inconsistencies, corrects source data if needed, and approves the validated dataset. After narrative generation, a human reviews the output against the data. The system doesn&#8217;t publish autonomously.</p><p><strong>Operational infrastructure.</strong> The pipeline tracks its own accuracy over time. Corrections feed back into the validation rules. The system versions its outputs so you can compare this month&#8217;s narratives against last month&#8217;s methodology.</p><p>Without the scaffold, this would be &#8220;use ChatGPT to write some paragraphs about KPIs.&#8221; With the scaffold, it&#8217;s a governed system that leadership trusts enough to present to the executive team.</p><div><hr></div><h2>The 3x multiplier</h2><p>The evidence that scaffolded approaches outperform ad hoc ones is strong.</p><p>Organizations with formal AI operating models achieve <a href="https://www.gcaie.org/post/the-ai-operating-model-playbook-redesigning-the-enterprise-for-scaled-ai">3.1x faster scaling and 4.4x higher value capture</a> compared to peers running ad hoc projects. They report 2.4x higher returns on AI investments. 72% of organizations successfully using AI at scale follow a defined operating framework.</p><p>The <a href="https://openai.com/index/the-state-of-enterprise-ai-2025-report/">OpenAI Enterprise report</a> adds another data point: purchasing AI tools from vendors and building partnerships succeeds 67% of the time, while internal builds succeed only 33%. The difference? Vendors bring pre-built operational scaffolds. You don&#8217;t start from scratch on input validation, guardrails, and monitoring - the vendor has already built it.</p><p>Frontier firms - the top quartile of AI adopters - show what operational integration looks like at scale. They achieve 1.7x revenue growth, 3.6x greater total shareholder return, and 1.6x EBIT margin compared to laggards. These aren&#8217;t cultural differences. They&#8217;re infrastructure differences. Frontier firms have built the scaffolds that let AI operate within business processes, not just alongside them.</p><p>Gartner predicts <a href="https://nexaitech.com/multi-ai-agent-architecutre-patterns-for-scale/">40% of agentic AI projects will be canceled by 2027</a> due to escalating costs, unclear business value, or inadequate risk controls. Read those failure modes: they&#8217;re all symptoms of missing scaffolding. Cost escalates because you retrofit governance after the fact. Value is unclear because you never defined how the AI connects to business outcomes. Risk controls are inadequate because you never designed them into the architecture.</p><p>The 60% that survive will be the ones that treated the scaffold as a first-class requirement, not an afterthought.</p><div><hr></div><h2>A fair objection: not everything needs a scaffold</h2><p>Some ad hoc AI usage genuinely works. <a href="https://www.nature.com/articles/s41562-026-02409-4">Research published in Nature Human Behaviour</a> shows that LLM-assisted writing improves outcomes even in informal settings - 18% of financial services complaints are now LLM-assisted, and 24% of corporate press releases are LLM-attributable.</p><p>This is real value, and it doesn&#8217;t require an operational scaffold. When a human writes a memo with ChatGPT&#8217;s help and then edits the result, the human is the scaffold. They validate the input (their own knowledge), apply decision rules (their judgment), standardize the output (their editing), and serve as the quality check.</p><p>The distinction is between <strong>utility</strong> and <strong>authority</strong>. When AI is a utility - humans decide, AI assists - light scaffolding works. When AI has authority - it acts, decides, or integrates into automated workflows - heavy scaffolding is required. The problem isn&#8217;t that organizations use passive AI. The problem is that they expect passive AI to deliver operational results.</p><div><hr></div><h2>What to do Monday morning</h2><p><strong>If you lead operations:</strong> Audit your AI pilots. How many are in production, processing real data, integrated into real workflows? If the answer is less than 15%, you have a scaffold problem, not a model problem. Fund the infrastructure, not another pilot.</p><p><strong>If you lead engineering:</strong> Stop thinking of AI as a feature you ship. Think of it as a system you operate. Design backwards from production constraints: what governs this AI? What validates its output? Who approves high-stakes decisions? What happens when it fails? Build observability into the initial design.</p><p><strong>If you lead ML:</strong> Your model is 10% of what production requires. The other 90% is input validation, decision rules, output standardization, human-in-the-loop workflows, and monitoring. Learn the frameworks - <a href="https://www.mckinsey.com/featured-insights/mckinsey-explainers/what-are-ai-guardrails/">McKinsey&#8217;s guardrails</a>, <a href="https://hai.stanford.edu/news/humans-loop-design-interactive-ai-systems">Stanford&#8217;s HITL patterns</a> - that turn models into operational systems. Partner with ops and engineering. You can&#8217;t build a scaffold alone.</p><p>The AI adoption race is largely over. Most enterprises have given their employees access to language models. The operational race - who can turn those models into governed, integrated, production-grade systems - is just starting. The scaffold is where it&#8217;s won.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://marcosnewsletter.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Marco's Substack is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[AI that doesn’t touch the P&L isn’t innovation, it’s overhead]]></title><description><![CDATA[Firms don&#8217;t need more experiments, they need workflows that raise margins.]]></description><link>https://marcosnewsletter.substack.com/p/ai-that-doesnt-touch-the-p-and-l</link><guid isPermaLink="false">https://marcosnewsletter.substack.com/p/ai-that-doesnt-touch-the-p-and-l</guid><dc:creator><![CDATA[Marco]]></dc:creator><pubDate>Fri, 26 Sep 2025 06:08:33 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!p4f6!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda17fa51-4913-421d-98f7-813279ab2b59_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>There&#8217;s a quiet frustration spreading across professional services firms. It&#8217;s not the technology itself. It&#8217;s what happens after the glossy demos, the pilots that win applause in town halls, the decks that promise &#8220;efficiency gains.&#8221;</p><p>The frustration shows up in the P&amp;L. Hours spent testing models that never reach production. Faster deliverables that paradoxically shrink fees. API invoices that grow without a clear link to client value. A portfolio of PoCs that look innovative on paper but, in reality, are margin erosion in disguise.</p><p>You can feel it in the Monday meetings. Leaders nod at &#8220;innovation updates&#8221; but privately wonder why gross margin hasn&#8217;t moved a single point. Teams are tired of context-switching between experiments while utilization targets remain stubborn. The real question isn&#8217;t <em>can AI do this?</em> It&#8217;s <em>should we even be doing this with AI at all?</em></p><div><hr></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!p4f6!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda17fa51-4913-421d-98f7-813279ab2b59_1536x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!p4f6!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda17fa51-4913-421d-98f7-813279ab2b59_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!p4f6!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda17fa51-4913-421d-98f7-813279ab2b59_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!p4f6!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda17fa51-4913-421d-98f7-813279ab2b59_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!p4f6!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda17fa51-4913-421d-98f7-813279ab2b59_1536x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!p4f6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda17fa51-4913-421d-98f7-813279ab2b59_1536x1024.png" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/da17fa51-4913-421d-98f7-813279ab2b59_1536x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2107705,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://marcosnewsletter.substack.com/i/172873027?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda17fa51-4913-421d-98f7-813279ab2b59_1536x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!p4f6!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda17fa51-4913-421d-98f7-813279ab2b59_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!p4f6!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda17fa51-4913-421d-98f7-813279ab2b59_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!p4f6!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda17fa51-4913-421d-98f7-813279ab2b59_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!p4f6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fda17fa51-4913-421d-98f7-813279ab2b59_1536x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>The pilots that never pay back</h2><p>The pain runs deeper than wasted hours. Consider how it manifests:</p><ul><li><p><strong>Decades of delivery playbooks get bypassed</strong> in the rush to launch &#8220;AI pilots&#8221; that aren&#8217;t scoped, priced, or risk-assessed.</p></li><li><p><strong>Client-facing deliverables risk errors</strong> that no one priced into the engagement, but everyone will be accountable for.</p></li><li><p><strong>Pressure from sales to &#8220;show AI&#8221;</strong> forces operators to demo half-baked use cases that consume margin instead of creating it.</p></li></ul><p>And yet, it feels lonely to say this out loud. Everyone else seems to be celebrating their pilots, right?</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://marcosnewsletter.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://marcosnewsletter.substack.com/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><h2>You are not the only one</h2><p>Here&#8217;s the truth: you&#8217;re far from alone.</p><p>BCG and Wharton recently ran a controlled experiment with 758 consultants. GPT-4 increased productivity by <strong>25%</strong> and improved quality on 18 tasks. But on &#8220;out-of-distribution&#8221; tasks, performance actually dropped. Translation: without proper gating and governance, AI doesn&#8217;t just fail to help   it actively hurts outcomes.</p><p>GitHub&#8217;s own Copilot study showed developers completing tasks <strong>55% faster</strong>. Yet firms that didn&#8217;t redesign their unit economics saw cycle times fall while fee structures stayed flat. Faster work simply compressed revenue.</p><p>Gartner projects that by 2026, <strong>80% of enterprises will run GenAI APIs in production</strong>. That means the competitive floor is rising quickly. The bar won&#8217;t be &#8220;do you use AI?&#8221; It will be &#8220;is your AI driving margin or bleeding it?&#8221;</p><p>The reality is simple: service firms that treat AI as a shiny pilot program will watch their margins shrink. Those that treat AI as a product line   with unit economics, SLOs, and governance will widen their edge</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!o-pu!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9897b5a2-6c5e-4f83-9a78-5d676acc0357_1024x1536.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!o-pu!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9897b5a2-6c5e-4f83-9a78-5d676acc0357_1024x1536.png 424w, https://substackcdn.com/image/fetch/$s_!o-pu!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9897b5a2-6c5e-4f83-9a78-5d676acc0357_1024x1536.png 848w, https://substackcdn.com/image/fetch/$s_!o-pu!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9897b5a2-6c5e-4f83-9a78-5d676acc0357_1024x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!o-pu!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9897b5a2-6c5e-4f83-9a78-5d676acc0357_1024x1536.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!o-pu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9897b5a2-6c5e-4f83-9a78-5d676acc0357_1024x1536.png" width="1024" height="1536" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9897b5a2-6c5e-4f83-9a78-5d676acc0357_1024x1536.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1536,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2666680,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://marcosnewsletter.substack.com/i/172873027?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9897b5a2-6c5e-4f83-9a78-5d676acc0357_1024x1536.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!o-pu!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9897b5a2-6c5e-4f83-9a78-5d676acc0357_1024x1536.png 424w, https://substackcdn.com/image/fetch/$s_!o-pu!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9897b5a2-6c5e-4f83-9a78-5d676acc0357_1024x1536.png 848w, https://substackcdn.com/image/fetch/$s_!o-pu!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9897b5a2-6c5e-4f83-9a78-5d676acc0357_1024x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!o-pu!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9897b5a2-6c5e-4f83-9a78-5d676acc0357_1024x1536.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>.</p><div><hr></div><h2>A different way forward: the TRIAGE map</h2><p>Instead of chasing every possible workflow, think about AI like an operator would: with constraints, priorities, and a hard eye on economics.</p><p>That&#8217;s where the <strong>TRIAGE framework</strong> comes in. It&#8217;s a way to select and industrialize use cases that actually move margins.</p><ul><li><p><strong>Throughput</strong>: How much cycle time or lead time could this cut?</p></li><li><p><strong>Risk</strong>: What happens if it fails? Is it client-facing (P0) or back-office (P2)? Do we have human-in-the-loop in place?</p></li><li><p><strong>Inputs</strong>: What data is needed? Do we have permission, residency, and confidentiality aligned?</p></li><li><p><strong>Accuracy</strong>: How do we measure quality? Golden sets, sampling, error thresholds?</p></li><li><p><strong>Governance</strong>: Are guardrails, audit logs, and approvals coded into the workflow?</p></li><li><p><strong>Economics</strong>: Does the cost of inference and integration improve realization rate, win rate, or pricing or cannibalize them?</p></li></ul><p>Once you score your workflows, you can literally plot them: Throughput vs Risk. The &#8220;quick wins&#8221; pop out high throughput, low risk. That&#8217;s where you should start.</p><p>This isn&#8217;t theory. Firms applying a structured filter like TRIAGE are moving from <strong>dozens of scattered pilots</strong> to <strong>five workflows that change margin profiles in 90 days</strong>. That&#8217;s the difference between &#8220;showing AI&#8221; and <em>selling AI</em>.</p><div><hr></div><h2>Why this matters right now</h2><p>Margins in services are under relentless pressure: flat contracts, rising delivery costs, and client expectations inflated by Microsoft and Google Copilot rollouts. The EU AI Act is adding compliance obligations. API costs are dropping, but that just tempts more experimentation without guardrails.</p><p>The worst outcome? Competitive parity. Everyone has AI, but no one makes money from it.</p><p>The best outcome? You design AI as a profit lever. Faster turnaround becomes a premium upsell, not a freebie. Enhanced accuracy becomes a differentiator, not an unbilled improvement. Guardrails and governance become a selling point, not a compliance headache.</p><div><hr></div><h2>What you can do this week</h2><p>Here&#8217;s a simple starting point: <strong>list ten recurring processes</strong> in your delivery engine. Score them on Throughput and Risk only. Plot them on a 2x2 grid. Circle the five that land in &#8220;high throughput, low risk.&#8221;</p><p>That&#8217;s your shortlist. Not thirty ideas. Not endless pilots. Just five.</p><p>Then and only then design SLOs, guardrails, and a shadow P&amp;L. Within two weeks, you&#8217;ll know if these workflows are margin-positive or margin-negative. Kill fast or scale fast. Nothing in between.</p><div><hr></div><h2>A practical step to try tomorrow</h2><p>In your next leadership meeting, skip the &#8220;AI update&#8221; slide. Instead, bring one number: <em>realization rate</em>. Ask, &#8220;Which AI workflows are actually improving this?&#8221; You&#8217;ll be amazed how quickly the noise falls away.</p><div><hr></div><h2>Let&#8217;s talk</h2><p>I&#8217;d love to hear from you:</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://marcosnewsletter.substack.com/p/ai-that-doesnt-touch-the-p-and-l/comments&quot;,&quot;text&quot;:&quot;Leave a comment&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://marcosnewsletter.substack.com/p/ai-that-doesnt-touch-the-p-and-l/comments"><span>Leave a comment</span></a></p><ul><li><p>Where have you seen AI pilots quietly erode margin instead of creating it?</p></li><li><p>Which workflows in your firm would score highest on Throughput and lowest on Risk right now?</p></li><li><p>If you had to kill 80% of your AI pilots tomorrow, which 20% would you keep   and why?</p></li></ul><p>Drop your thoughts in the comments. The sharper, the better. The real value is in seeing how others are navigating this same map.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://marcosnewsletter.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption"></p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Your “Strategy” Is Just a To-Do List (and It’s Costing You Growth)]]></title><description><![CDATA[Pretty roadmaps won&#8217;t save you. Outcomes do. Here&#8217;s how to tell if you&#8217;re running on strategy or just a glorified to-do list.]]></description><link>https://marcosnewsletter.substack.com/p/your-strategy-is-just-a-to-do-list</link><guid isPermaLink="false">https://marcosnewsletter.substack.com/p/your-strategy-is-just-a-to-do-list</guid><dc:creator><![CDATA[Marco]]></dc:creator><pubDate>Wed, 24 Sep 2025 06:08:25 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!2c1Q!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d481833-c441-4272-8da3-f66b2066747e_864x484.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Let me start with something positive: every team I&#8217;ve worked with is full of people who genuinely want to do good work. No one wakes up hoping to create more bureaucracy, more slide decks, or more noise. They want impact. They want clarity. They want to move the ball forward.</p><p>And yet every week I see the same silent killer inside organizations: <strong>teams confuse a plan with a strategy.</strong></p><div><hr></div><h2>The moment it all blurs</h2><p>It usually starts in a meeting room (or more likely, a Zoom call). Someone asks: <em>&#8220;What&#8217;s our strategy?&#8221;</em></p><p>What follows isn&#8217;t strategy it&#8217;s a Gantt chart. A backlog. A roadmap.</p><p>In other words: a plan.</p><p>On paper, the distinction looks obvious. <strong>Strategy is the &#8220;why&#8221; and &#8220;where to play.&#8221; A plan is the &#8220;how&#8221; and &#8220;when.&#8221;</strong> But in reality, the two get mixed all the time and the cost is high.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://marcosnewsletter.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption"></p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><h2>The pain of mistaking a plan for a strategy</h2><p>If you&#8217;ve ever been in one of these situations, you know the feeling:</p><ul><li><p>A &#8220;strategy deck&#8221; that&#8217;s really just a 40-slide list of initiatives.</p></li><li><p>Teams pivoting direction mid-quarter because the latest request feels urgent.</p></li><li><p>Deliverables shipped on time, but with no noticeable change in customer behavior or revenue.</p></li><li><p>Endless meetings where leaders debate tasks, not outcomes.</p></li></ul><p>It&#8217;s exhausting. It&#8217;s confusing. And most importantly, it&#8217;s wasteful.</p><p>PwC found that <strong>67% of managers confuse operational planning with strategy.</strong> That means two out of three managers are pushing their teams forward with what amounts to <em>organized noise.</em></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2c1Q!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d481833-c441-4272-8da3-f66b2066747e_864x484.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2c1Q!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d481833-c441-4272-8da3-f66b2066747e_864x484.webp 424w, https://substackcdn.com/image/fetch/$s_!2c1Q!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d481833-c441-4272-8da3-f66b2066747e_864x484.webp 848w, https://substackcdn.com/image/fetch/$s_!2c1Q!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d481833-c441-4272-8da3-f66b2066747e_864x484.webp 1272w, https://substackcdn.com/image/fetch/$s_!2c1Q!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d481833-c441-4272-8da3-f66b2066747e_864x484.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2c1Q!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d481833-c441-4272-8da3-f66b2066747e_864x484.webp" width="864" height="484" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0d481833-c441-4272-8da3-f66b2066747e_864x484.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:484,&quot;width&quot;:864,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:22690,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/webp&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://marcosnewsletter.substack.com/i/171987571?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d481833-c441-4272-8da3-f66b2066747e_864x484.webp&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!2c1Q!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d481833-c441-4272-8da3-f66b2066747e_864x484.webp 424w, https://substackcdn.com/image/fetch/$s_!2c1Q!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d481833-c441-4272-8da3-f66b2066747e_864x484.webp 848w, https://substackcdn.com/image/fetch/$s_!2c1Q!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d481833-c441-4272-8da3-f66b2066747e_864x484.webp 1272w, https://substackcdn.com/image/fetch/$s_!2c1Q!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0d481833-c441-4272-8da3-f66b2066747e_864x484.webp 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><h2>You&#8217;re not alone</h2><p>This isn&#8217;t a rare mistake it&#8217;s systemic.</p><ul><li><p>McKinsey data shows companies that tightly connect strategy and planning are <strong>3x more likely to outperform competitors</strong> on revenue growth.</p></li><li><p>Bain found that organizations focusing on outcomes (customer retention, churn reduction, market share) have a <strong>2.5x higher project success rate</strong> than those tracking outputs (deliverables, task completion).</p></li><li><p>Gartner predicts that by 2027, <strong>70% of operational planning will be automated by AI.</strong> Which means the last real human edge will be in strategy: choosing <em>why</em> and <em>where</em> not <em>how</em>.</p></li></ul><p>When you zoom out, the picture becomes clear: companies don&#8217;t fail because they lack plans. They fail because their plans aren&#8217;t anchored to a strategy.</p><div><hr></div><h2>Common traps I see every day</h2><p>The same mistakes repeat across industries:</p><ul><li><p><strong>Turning vision into activity.</strong> A bold idea gets diluted into a laundry list of initiatives, none of which connect to the core purpose.</p></li><li><p><strong>Adding detail without power.</strong> Teams think more detail = more control, but what they&#8217;re really doing is burying the few decisions that matter.</p></li><li><p><strong>Planning in silos.</strong> Each department builds its own plan, with no view of dependencies or shared goals.</p></li></ul><p>The end result? Momentum without meaning. Lots of movement, little impact.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!QBYO!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa4c160fe-311c-492c-8a7c-7aa98b367ed7_766x430.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!QBYO!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa4c160fe-311c-492c-8a7c-7aa98b367ed7_766x430.webp 424w, https://substackcdn.com/image/fetch/$s_!QBYO!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa4c160fe-311c-492c-8a7c-7aa98b367ed7_766x430.webp 848w, https://substackcdn.com/image/fetch/$s_!QBYO!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa4c160fe-311c-492c-8a7c-7aa98b367ed7_766x430.webp 1272w, https://substackcdn.com/image/fetch/$s_!QBYO!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa4c160fe-311c-492c-8a7c-7aa98b367ed7_766x430.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!QBYO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa4c160fe-311c-492c-8a7c-7aa98b367ed7_766x430.webp" width="766" height="430" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a4c160fe-311c-492c-8a7c-7aa98b367ed7_766x430.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:430,&quot;width&quot;:766,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:15300,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/webp&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://marcosnewsletter.substack.com/i/171987571?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa4c160fe-311c-492c-8a7c-7aa98b367ed7_766x430.webp&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!QBYO!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa4c160fe-311c-492c-8a7c-7aa98b367ed7_766x430.webp 424w, https://substackcdn.com/image/fetch/$s_!QBYO!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa4c160fe-311c-492c-8a7c-7aa98b367ed7_766x430.webp 848w, https://substackcdn.com/image/fetch/$s_!QBYO!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa4c160fe-311c-492c-8a7c-7aa98b367ed7_766x430.webp 1272w, https://substackcdn.com/image/fetch/$s_!QBYO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa4c160fe-311c-492c-8a7c-7aa98b367ed7_766x430.webp 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><h2>What actually works</h2><p>Here&#8217;s the shift I&#8217;ve seen in teams that get this right:</p><ol><li><p><strong>They ask &#8220;why&#8221; before &#8220;how.&#8221;</strong><br>If the only answer to &#8220;why are we doing this?&#8221; is &#8220;because it&#8217;s in the roadmap,&#8221; you&#8217;re in trouble. Outcomes must drive activities, not the other way around.</p></li><li><p><strong>They keep strategy short.</strong><br>No 40-slide decks. Two pages max. If you can&#8217;t explain your strategy over coffee, no plan will save you.</p></li><li><p><strong>They measure what matters.</strong><br>Activity metrics are fine for daily management, but what moves the business are outcome metrics cycle times shrinking, margins improving, customer loyalty rising.</p></li></ol><p>These teams don&#8217;t have less planning they have <em>better anchored planning.</em></p><div><hr></div><h2>The checklist test</h2><p>If you&#8217;re wondering whether you&#8217;re really working on a strategy or just a plan, here&#8217;s a quick test. Run your work against these five questions:</p><ol><li><p><strong>Clarity</strong> &#8211; Can you explain your strategy in two pages or less?</p></li><li><p><strong>Focus</strong> &#8211; Have you clearly said <em>what not to do</em>?</p></li><li><p><strong>Connection</strong> &#8211; Do your plans tie directly back to strategic outcomes?</p></li><li><p><strong>Measurement</strong> &#8211; Are you tracking outcomes (impact, results) instead of just outputs (tasks, deliverables)?</p></li><li><p><strong>Review</strong> &#8211; Do you revisit the strategy regularly without rewriting it from scratch every time?</p></li></ol><p>If you can&#8217;t check most of these boxes, chances are you&#8217;re running on a plan not a strategy.</p><div><hr></div><h2>A practical move for tomorrow</h2><p>Here&#8217;s something you can try right away: <strong>for your next project kickoff, ban the words &#8220;deliverable&#8221; and &#8220;timeline&#8221; from the first 20 minutes.</strong></p><p>Instead, ask: <em>&#8220;What will be different if this project succeeds?&#8221;</em></p><p>If the answer is still just &#8220;we&#8217;ll ship X feature by Y date,&#8221; go one level deeper. Push until you hit a customer behavior, a business metric, or a competitive position. Only then start talking about the plan.</p><div><hr></div><h2>Closing on a positive note</h2><p>Here&#8217;s the good news: this isn&#8217;t rocket science. Most teams already have the raw ingredients for a real strategy they just bury it under activity. When you bring the focus back to outcomes, you unlock energy. Conversations shift from tasks to impact. People stop feeling like they&#8217;re chasing deadlines and start feeling like they&#8217;re building something that matters.</p><p>And that&#8217;s what makes work worth doing.</p><div><hr></div><h3>Let&#8217;s talk</h3><p>Now I want to throw the question back to you:</p><ul><li><p>In your daily work, do you see <strong>more plans with no strategy, or strategies that never make it into a plan</strong>?</p></li><li><p>What&#8217;s the hardest part of keeping those two connected in your organization?</p></li><li><p>And if you could change just one thing tomorrow to strengthen that connection, what would it be?</p></li></ul><p>I&#8217;d love to hear your take drop a comment and let&#8217;s spark the kind of conversation that moves this from theory into practice.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://marcosnewsletter.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Marco, Future&#8209;Ops Insights is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[RPA, AI, or Agents? The Messy Truth About Choosing Automation]]></title><description><![CDATA[The hardest part isn&#8217;t the technology. It&#8217;s knowing which problem you&#8217;re really trying to solve.]]></description><link>https://marcosnewsletter.substack.com/p/rpa-ai-or-agents-the-messy-truth</link><guid isPermaLink="false">https://marcosnewsletter.substack.com/p/rpa-ai-or-agents-the-messy-truth</guid><dc:creator><![CDATA[Marco]]></dc:creator><pubDate>Mon, 22 Sep 2025 06:08:36 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!LCri!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77d61d02-dd6d-4150-8fda-b188a55d6f00_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>It always starts the same way. A team is drowning in repetitive work, leaders decide it&#8217;s time to &#8220;automate,&#8221; and suddenly everyone has a different definition of what that means. Some picture rule-based bots cranking through invoices. Others imagine sleek AI systems predicting customer churn. A few are already whispering about autonomous agents that can run entire workflows on their own.</p><p>And then comes the paralysis. Which one do we actually need? What&#8217;s hype, what&#8217;s real, and what will quietly collapse after a year of maintenance headaches?</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!LCri!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77d61d02-dd6d-4150-8fda-b188a55d6f00_1536x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!LCri!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77d61d02-dd6d-4150-8fda-b188a55d6f00_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!LCri!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77d61d02-dd6d-4150-8fda-b188a55d6f00_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!LCri!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77d61d02-dd6d-4150-8fda-b188a55d6f00_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!LCri!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77d61d02-dd6d-4150-8fda-b188a55d6f00_1536x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!LCri!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77d61d02-dd6d-4150-8fda-b188a55d6f00_1536x1024.png" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/77d61d02-dd6d-4150-8fda-b188a55d6f00_1536x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2411644,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://marcosnewsletter.substack.com/i/172564033?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77d61d02-dd6d-4150-8fda-b188a55d6f00_1536x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!LCri!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77d61d02-dd6d-4150-8fda-b188a55d6f00_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!LCri!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77d61d02-dd6d-4150-8fda-b188a55d6f00_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!LCri!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77d61d02-dd6d-4150-8fda-b188a55d6f00_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!LCri!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F77d61d02-dd6d-4150-8fda-b188a55d6f00_1536x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><h2>The Pain of Picking the Wrong Tool</h2><p>Many organizations jump straight into automation projects with the wrong match of tool to problem. The fallout is costly.</p><ul><li><p><strong>Rigid RPA bots</strong> break when even a minor system update shifts a field in the interface. Teams spend more time fixing bots than they save.</p></li><li><p><strong>AI pilots</strong> end up stalled because no one prepared the right data, and the models give outputs that look impressive but don&#8217;t actually help anyone make better decisions.</p></li><li><p><strong>Autonomous agents</strong> promise freedom but in practice spin into loops or generate answers that no one can trust without heavy supervision.</p></li></ul><p>The result: leaders who invested millions are left asking why the promised efficiency gains never materialized. Employees feel frustrated they wanted help, not another layer of complexity.</p><div><hr></div><h2>You&#8217;re Not Alone in This</h2><p>If this sounds familiar, it&#8217;s because it is. According to Deloitte, around <strong>70% of companies exploring automation now seek more flexible and autonomous solutions</strong> after realizing the limitations of rule-based approaches. Meanwhile, Accenture reports that <strong>55% of developers using GitHub Copilot finished coding tasks faster</strong>, but many still relied on human review to catch mistakes.</p><p>Even at the cutting edge, early adopters admit reality doesn&#8217;t always match the pitch. In 2024, pilot projects with autonomous agents showed promise think virtual junior consultants drafting reports or IT agents fixing routine issues but they also revealed brittleness, hallucinations, and a clear need for human oversight.</p><p>So if your own projects feel messy or inconclusive, know this: it&#8217;s not a personal failure. It&#8217;s the stage the entire industry is in.</p><div><hr></div><h2>The Real Question: What Problem Are You Solving?</h2><p>The temptation is to treat &#8220;automation&#8221; as a single bucket. In practice, each category <strong>traditional automation, AI, and autonomous agents</strong> solves very different problems.</p><ul><li><p><strong>Traditional automation (RPA, scripts):</strong> Think of it as the specialist for <strong>repetitive, structured, and high-volume tasks</strong>. It thrives on invoices, reconciliations, and copy-paste operations between systems. When stability and predictability matter, RPA is unbeatable. But once a process requires judgment or adaptation, it falls apart.</p></li><li><p><strong>AI (machine learning, NLP, predictive models):</strong> This is the analyst, the pattern-recognizer, the one that makes sense of <strong>messy or unstructured data</strong>. It&#8217;s how law firms like Luminance cut contract review time by 60%, and how consultancies forecast market shifts faster. The caveat: AI is only as strong as the data behind it, and bias or poor governance can derail results.</p></li><li><p><strong>Autonomous agents:</strong> These are still the apprentices, learning fast but not yet fully trusted. They shine in <strong>complex, multi-step workflows</strong> where flexibility is essential like diagnosing tech issues end-to-end or coordinating compliance checks across thousands of documents. But think of them like a promising intern: smart, fast, but still needing a mentor&#8217;s eye.</p></li></ul><p>When you frame the choice this way, the fog clears. It&#8217;s less about which technology is &#8220;best&#8221; and more about which one fits the shape of your problem.</p><div><hr></div><h2>Why Mixing Matters</h2><p>The future isn&#8217;t one or the other it&#8217;s all of them, working together. Some of the most impressive results already come from <strong>hybrid approaches</strong>.</p><p>A US health insurer tripled document processing by combining bots for routine intake with AI models for flagging ambiguous cases. A global credit card company saved <strong>$160 million</strong> by blending RPA for payables with AI for fraud detection. And in professional services, firms are layering chatbots for client FAQs on top of AI-powered analytics, with humans handling the final review.</p><p>It&#8217;s like building a team: you wouldn&#8217;t hire only accountants or only strategists. You need the right mix of skills for the job. Automation is no different.</p><div><hr></div><h2>A Practical Way Forward</h2><p>Here&#8217;s where optimism meets pragmatism: the solution is rarely about chasing the newest tool. It&#8217;s about starting with a clear problem definition, then matching it to the simplest technology that solves it.</p><ul><li><p>If your process is stable, repetitive, and structured <strong>start with RPA</strong>.</p></li><li><p>If the challenge lies in patterns, predictions, or messy data <strong>AI is your friend</strong>.</p></li><li><p>If the problem is complex, dynamic, and can&#8217;t be easily scripted <strong>experiment with agents</strong>, but keep humans in the loop.</p></li></ul><p>This doesn&#8217;t just save money. It builds confidence across your team. People see quick wins from automation, trust grows, and the organization is more willing to experiment with advanced tools down the line.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://marcosnewsletter.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://marcosnewsletter.substack.com/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><h2>One Action You Can Take This Week</h2><p>Instead of debating abstractly which automation to pursue, pick <strong>one process in your team that everyone complains about</strong>. Map it out on a whiteboard. Ask: is it structured and repetitive, data-driven and messy, or complex and dynamic? That single classification will point you toward whether it needs RPA, AI, or an agent experiment.</p><p>It&#8217;s not about solving everything at once. It&#8217;s about getting one match right and proving to your team that automation can work <em>for</em> them, not against them.</p><div><hr></div><h2>Let&#8217;s Talk</h2><p>What&#8217;s the one process in your organization that drains the most energy?<br>Have you ever tried to automate it and if so, which approach did you choose?<br>Where did it succeed, and where did it break?</p><p>I&#8217;d love to hear your stories. Drop them in the comments not just the wins, but the failures too. Those are the ones that usually teach us the most.</p><h1>Full Detailed Report on + Executive Summary at the end:</h1><h1>Comparing AI Deployment Models: On-Premises Hardware vs Cloud vs Turnkey Solutions</h1><p>Artificial Intelligence can be deployed in several ways, each with its own benefits and adoption patterns. The three primary models are <strong>AI on physical hardware (on-premises)</strong>, <strong>AI on the cloud</strong>, and <strong>turnkey AI solutions</strong>. Below we explore each model in depth, including recent global trends (focused on the past year) and industry usage statistics.</p>
      <p>
          <a href="https://marcosnewsletter.substack.com/p/rpa-ai-or-agents-the-messy-truth">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Should You Own or Rent Your AI? The Hidden Costs Behind Hardware, Cloud, and ChatGPT]]></title><description><![CDATA[The honest math behind control, speed, and cost in the AI era]]></description><link>https://marcosnewsletter.substack.com/p/should-you-own-or-rent-your-ai-the</link><guid isPermaLink="false">https://marcosnewsletter.substack.com/p/should-you-own-or-rent-your-ai-the</guid><dc:creator><![CDATA[Marco]]></dc:creator><pubDate>Mon, 15 Sep 2025 06:08:32 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!rR_h!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ce4a00d-63af-4350-91ee-5ddef113eea9_1024x1536.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Let&#8217;s start on a positive note: it&#8217;s an extraordinary time to be alive if you&#8217;re curious about AI. For the first time, powerful models that once required research labs and million-dollar budgets are now accessible to startups, mid-sized firms, even solo entrepreneurs. You can spin up a model in the cloud within minutes, or if you&#8217;re ambitious download one onto a spare GPU at home. The possibilities are wide open.</p><p>But with possibility comes friction. And that&#8217;s where the story begins.</p><div><hr></div><h2>The Problem Nobody Likes to Admit</h2><p>We&#8217;ve all felt that pang of doubt when choosing between building in-house versus relying on a service. With AI, the stakes feel even higher.</p><p>Imagine this: your team wants to experiment with generative models. Marketing sees the chance to generate on-brand content instantly. Operations dreams of smarter forecasting. Legal, meanwhile, is pacing in the corner, whispering &#8220;GDPR&#8221; under their breath.</p><p>The enthusiasm is real but so is the paralysis. <strong>Do you spend six figures on GPUs, hire engineers, and pray your hardware doesn&#8217;t become obsolete in two years? Or do you rent elastic power from AWS and risk waking up one morning with a bill that looks like the GDP of a small town? Or do you just swipe your credit card for ChatGPT Plus and accept the trade-offs?</strong></p><p>Each option feels like both an opportunity and a trap.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!rR_h!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ce4a00d-63af-4350-91ee-5ddef113eea9_1024x1536.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!rR_h!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ce4a00d-63af-4350-91ee-5ddef113eea9_1024x1536.png 424w, https://substackcdn.com/image/fetch/$s_!rR_h!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ce4a00d-63af-4350-91ee-5ddef113eea9_1024x1536.png 848w, https://substackcdn.com/image/fetch/$s_!rR_h!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ce4a00d-63af-4350-91ee-5ddef113eea9_1024x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!rR_h!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ce4a00d-63af-4350-91ee-5ddef113eea9_1024x1536.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!rR_h!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ce4a00d-63af-4350-91ee-5ddef113eea9_1024x1536.png" width="1024" height="1536" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5ce4a00d-63af-4350-91ee-5ddef113eea9_1024x1536.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1536,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1802797,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://marcosnewsletter.substack.com/i/172562100?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ce4a00d-63af-4350-91ee-5ddef113eea9_1024x1536.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!rR_h!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ce4a00d-63af-4350-91ee-5ddef113eea9_1024x1536.png 424w, https://substackcdn.com/image/fetch/$s_!rR_h!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ce4a00d-63af-4350-91ee-5ddef113eea9_1024x1536.png 848w, https://substackcdn.com/image/fetch/$s_!rR_h!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ce4a00d-63af-4350-91ee-5ddef113eea9_1024x1536.png 1272w, https://substackcdn.com/image/fetch/$s_!rR_h!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ce4a00d-63af-4350-91ee-5ddef113eea9_1024x1536.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><h2>Pain Points in Sharp Relief</h2><p>The more you dig, the more the anxiety builds.</p><ul><li><p><strong>Hardware lock-in</strong>: Buying servers feels bold until you realize NVIDIA has already launched the &#8220;next&#8221; card three months later.</p></li><li><p><strong>Cloud sticker shock</strong>: Those $2/hour GPU rentals look harmless until you realize your team left three instances running over the weekend.</p></li><li><p><strong>Data worries</strong>: You can&#8217;t shake the feeling that pushing client health records or banking logs into a third-party API is asking for tomorrow&#8217;s compliance nightmare.</p></li><li><p><strong>Skill gaps</strong>: You may have brilliant analysts, but who in your company is actually capable of fine-tuning a 70B parameter model?</p></li></ul><p>What makes it worse is that you&#8217;re not alone in this paralysis. A 2024 Deloitte survey found <strong>53% of executives say AI adoption is slowed primarily by &#8220;infrastructure uncertainty.&#8221;</strong> Not algorithms. Not talent. Infrastructure.</p><p>That&#8217;s telling. The bottleneck isn&#8217;t whether AI works it&#8217;s where it should live.</p><div><hr></div><h2>Stories from the Field</h2><ul><li><p><strong>The fintech startup drowning in cloud bills</strong><br>FeroTrize a venture-backed fintech startup delivering real-time lending services at scale, built on a robust AWS-native architecture. With rapid market adoption, the company was processing thousands of transactions per minute. But as innovation surged, so did their AWS bill ballooning from $15,000 to $48,000 per month in just 18 months. As their CTO Marcus Chen admitted: <em>&#8220;We were focused on rapid growth and feature development, but our AWS bills were growing even faster than our customer base.&#8221;</em> <em><a href="https://www.stackgenie.io/how-stackgenie-helped-a-fintech-startup-cut-aws-cloud-costs-by-40-with-eks/?utm_source=chatgpt.com">Source</a></em></p></li><li><p><strong>German hospitals resisting the cloud</strong><br><strong>Asklepios Kliniken</strong>, one of the country&#8217;s largest private hospital groups, rolled out AI company <strong>Aidoc&#8217;s aiOS&#8482; platform</strong> across <strong>25 hospitals</strong>, but deployed it in a way that kept control local to ensure data sovereignty. They spent &#8364;800,000 building the on-premise AI cluster for patient data analysis. Twelve months later, the GPUs were maxed out, and they had to rent additional cloud compute anyway. <em><a href="https://www.aidoc.com/about/news/asklepios-and-aidoc-set-new-standard-for-patient-care/?utm_source=chatgpt.com">Source</a></em></p></li></ul><p>Different choices, different consequences. Both path are messy, but it&#8217;s important to see that companies of all sizes are making these bets and adjusting along the way.</p><div><hr></div><h2>The Escape: Clarity on Trade-Offs</h2><p>The good news? There isn&#8217;t a single &#8220;right&#8221; answer. There are just <strong>trade-offs</strong> and clarity about those trade-offs is liberating.</p><ul><li><p><strong>On-premise</strong> gives you sovereignty. You own the stack, the data never leaves your walls, and you can customize deeply. <strong>The trade-off</strong>: upfront cost and maintenance headaches.</p></li><li><p><strong>Cloud</strong> gives you speed and scale. You rent world-class GPUs without tying up capital. <strong>The trade-off:</strong> recurring costs and dependence on a provider&#8217;s rules.</p></li><li><p><strong>As-a-Service (ChatGPT, Claude, Gemini)</strong> gives you instant productivity. No DevOps, no capital expense. <strong>The trade-off</strong>: zero control over the model, and limited data privacy.</p></li></ul><p>Seen clearly, the question stops being &#8220;what&#8217;s best?&#8221; and becomes &#8220;what&#8217;s best for me right now?&#8221;</p><div><hr></div><h2>A Framework That Actually Helps</h2><p>When you strip away the noise, a practical lens emerges:</p><ol><li><p><strong>If you&#8217;re prototyping or solo</strong> &#8594; SaaS models (ChatGPT, Claude) make sense. Your risk is low, and your speed is high.</p></li><li><p><strong>If you&#8217;re scaling a product where AI is the core</strong> &#8594; Cloud GPUs plus open-source models like LLaMA 3 or Mistral are your friend. You can scale elastically while keeping some customization.</p></li><li><p><strong>If your data is mission-critical (healthcare, finance, defense)</strong> &#8594; On-premise is still the fortress. It&#8217;s expensive, yes, but sovereignty often is.</p></li></ol><p>This framework doesn&#8217;t eliminate trade-offs, but it gives you a starting map. And a map is better than wandering blind.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://marcosnewsletter.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://marcosnewsletter.substack.com/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><h2>A Practical Note</h2><p>If you&#8217;re standing at this crossroads right now, here&#8217;s one small but powerful step you can take this week:</p><p><strong>Run a pilot project with a single use case, using two different infrastructures.</strong></p><p>For example: try a customer-support summarizer on ChatGPT, and simultaneously test a fine-tuned open-source model on rented cloud GPUs. Track cost, latency, and output quality for just one week. You&#8217;ll learn more from that side-by-side than from any whitepaper.</p><p>That&#8217;s your low-risk way to feel the trade-offs in your bones not just in a slide deck.</p><div><hr></div><p>I&#8217;ll leave you with this thought: we often overcomplicate &#8220;strategy&#8221; when a cheap, fast experiment can cut through the fog.</p><p>Now I&#8217;d love to hear from you:<br>&#128073; <strong>If you had to bet on one setup today hardware, cloud, or SaaS where would you place your chips, and why?</strong><br>&#128073; <strong>What&#8217;s the single biggest fear holding you back from committing?</strong></p><p>Drop your thoughts in the comments I want to surface the real stories, not just the glossy case studies.</p><h1>Full Report on: AI, automation, and autonomous agents in professional services: when to use what</h1><p>In professional services, technologies like traditional automation, artificial intelligence, and autonomous agents are reshaping operations. They promise higher efficiency, lower costs, and better quality, but each is suited to different scenarios. The real advantage comes from knowing when to use rule-based automation, when to rely on AI and machine learning, and when to experiment with autonomous agents. Many firms still win with automation on repetitive tasks, while a growing share is piloting more flexible, autonomous approaches to stay competitive. Below is a practical, research-informed overview of each category, with guidance on when to use it, common use cases, and real examples aligned to services operations.</p>
      <p>
          <a href="https://marcosnewsletter.substack.com/p/should-you-own-or-rent-your-ai-the">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[AI isn’t the strategy. Leadership is.]]></title><description><![CDATA[The firms pulling ahead in AI share four mandates. Miss them, and you&#8217;re funding experiments. Nail them, and you&#8217;re compounding advantage.]]></description><link>https://marcosnewsletter.substack.com/p/ai-isnt-the-strategy-leadership-is</link><guid isPermaLink="false">https://marcosnewsletter.substack.com/p/ai-isnt-the-strategy-leadership-is</guid><dc:creator><![CDATA[Marco]]></dc:creator><pubDate>Fri, 12 Sep 2025 06:08:20 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!yHd3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F809378b5-7be8-4fac-ae28-3179b4a93109_1536x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Let&#8217;s be honest: <strong>AI in business is at risk of becoming the next &#8220;digital transformation&#8221; clich&#233;.</strong> Everyone talks about it, budgets are being poured in, but when you look closely, most initiatives are still sitting in pilot mode or, worse, producing little more than shiny slide decks.</p><p>If you&#8217;ve felt the frustration of sinking time and resources into AI only to watch outcomes stall, you&#8217;re not alone. A recent Gartner study found that <strong>over 80% of AI projects never make it past experimentation.</strong> That&#8217;s not a technology failure it&#8217;s a leadership one.</p><p>And that&#8217;s the real problem: executives are buying tools, but not building the conditions for transformation.</p><div><hr></div><h2>The Pain Point Nobody Likes to Admit</h2><p>Picture this. A boardroom presentation announces the company&#8217;s &#8220;AI journey.&#8221; The team proudly showcases a chatbot prototype, a few automations in the back office, and some dashboards that promise predictive insights.</p><p>Then, six months later, the results are flat. Sales hasn&#8217;t grown, costs haven&#8217;t gone down, and customer satisfaction hasn&#8217;t budged. Everyone involved feels deflated. Some even whisper that AI is overhyped.</p><p>The truth? The organization didn&#8217;t fail because of the AI. It failed because leadership treated AI as a <strong>tech decision</strong> not a <strong>strategic mandate.</strong></p><p>This is the pain point: too many leaders are delegating AI to IT or innovation labs without anchoring it to business outcomes. It&#8217;s like installing solar panels on a collapsing roof the technology is fine, but the foundation isn&#8217;t there.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!yHd3!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F809378b5-7be8-4fac-ae28-3179b4a93109_1536x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!yHd3!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F809378b5-7be8-4fac-ae28-3179b4a93109_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!yHd3!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F809378b5-7be8-4fac-ae28-3179b4a93109_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!yHd3!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F809378b5-7be8-4fac-ae28-3179b4a93109_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!yHd3!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F809378b5-7be8-4fac-ae28-3179b4a93109_1536x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!yHd3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F809378b5-7be8-4fac-ae28-3179b4a93109_1536x1024.png" width="1456" height="971" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/809378b5-7be8-4fac-ae28-3179b4a93109_1536x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:971,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2557797,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://marcosnewsletter.substack.com/i/172156491?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F809378b5-7be8-4fac-ae28-3179b4a93109_1536x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!yHd3!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F809378b5-7be8-4fac-ae28-3179b4a93109_1536x1024.png 424w, https://substackcdn.com/image/fetch/$s_!yHd3!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F809378b5-7be8-4fac-ae28-3179b4a93109_1536x1024.png 848w, https://substackcdn.com/image/fetch/$s_!yHd3!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F809378b5-7be8-4fac-ae28-3179b4a93109_1536x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!yHd3!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F809378b5-7be8-4fac-ae28-3179b4a93109_1536x1024.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><h2>Proof That You&#8217;re Not Alone</h2><p>This pattern repeats everywhere. McKinsey reports that while <strong>65% of companies say they&#8217;ve adopted AI in some form, only 23% report measurable ROI.</strong> In other words, most firms are dabbling, but very few are capturing value.</p><p>Even among Fortune 500 companies, success is uneven. Some think JPMorgan or Walmart are achieving billions in efficiency and revenue gains. Others are spinning their wheels, unable to get past fragmented pilots.</p><p>The gap comes down to leadership, not technology. The winners have <strong>clear vision, strong foundations, empowered people, and disciplined execution.</strong> The laggards don&#8217;t.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://marcosnewsletter.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://marcosnewsletter.substack.com/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><h2>Four Leadership Mandates</h2><p>The good news? The pattern is clear enough to learn from. Across industries, the companies breaking through with AI share four common leadership actions.</p><div><hr></div><h3>1. Lead with a Clear, Top-Down Vision</h3><p>The number one reason AI projects die is lack of vision. Without a north star, efforts fragment into disconnected experiments.</p><p>Successful leaders define the &#8220;why&#8221; up front: <em>Are we using AI to cut cycle times in half? To expand market share by 30%? To reduce customer churn by 15%?</em></p><p>And they articulate it relentlessly. When Satya Nadella reframed Microsoft as an &#8220;AI-first&#8221; company, it wasn&#8217;t a slogan. It became a mandate guiding everything from product to hiring to partnerships.</p><p><strong>Your move:</strong> If you can&#8217;t explain in one sentence how AI ties directly to your company&#8217;s competitive advantage, you don&#8217;t have a strategy you have a hobby.</p><div><hr></div><h3>2. Build the Foundations: Data and Integration</h3><p>Here&#8217;s the harsh reality: AI is only as good as the data it learns from. And most companies have messy, siloed, low-integrity data.</p><p>Deploying sophisticated AI on top of bad data doesn&#8217;t just fail it creates <strong>sophisticated failures.</strong></p><p>That&#8217;s why leading firms invest heavily in centralized, clean, and integrated data ecosystems before scaling AI. They also avoid point-solution chaos by building platforms with open APIs, so systems can talk to each other.</p><p>Amazon didn&#8217;t dominate because it had the best algorithms it dominated because it had the most complete, unified data about customers.</p><p><strong>Your move:</strong> Don&#8217;t just ask &#8220;What AI tool should we buy?&#8221; Ask &#8220;Do we have the foundation for AI to actually work?&#8221;</p><div><hr></div><h3>3. Invest in an AI-Ready Workforce and Culture</h3><p>AI doesn&#8217;t replace people it <strong>reshapes their roles.</strong> But if your team sees AI as a threat, adoption dies before it begins.</p><p>The companies pulling ahead are making AI literacy a baseline skill, not a niche one. According to PwC, employees who feel AI-fluent report <strong>96% higher productivity and job satisfaction.</strong></p><p>Even more important, these organizations create <strong>psychological safety</strong> the freedom to reinvent processes without fear of failure. Middle managers aren&#8217;t told to &#8220;optimize&#8221; old workflows; they&#8217;re encouraged to design entirely new ones.</p><p><strong>Your move:</strong> Train your workforce not just to <em>use</em> AI, but to <em>think</em> AI.</p><div><hr></div><h3>4. Execute in Phases, Tied to ROI</h3><p>The final mandate is discipline. Too many leaders chase &#8220;big bang&#8221; transformations. They throw millions at AI and hope scale will equal success.</p><p>It rarely does. The leaders who succeed take a <strong>phased, ROI-driven approach.</strong></p><p>They start with one high-impact pilot: &#8220;Increase win rate from 22% to 25% in 90 days.&#8221; When that works, they expand. Each phase validates the next.</p><p>It&#8217;s not glamorous, but it&#8217;s effective.</p><p><strong>Your move:</strong> Anchor every AI initiative to a measurable outcome. If you can&#8217;t define the metric, don&#8217;t fund the project.</p><div><hr></div><h2>Why This Matters Now</h2><p>The shift to AI-native organizations isn&#8217;t a distant trend it&#8217;s already reshaping markets. Firms that build these four leadership muscles will widen the gap fast.</p><p>Those that don&#8217;t? They&#8217;ll keep spending money on AI while watching competitors compound advantages. And because AI compounds exponentially, the window to catch up is shorter than most leaders realize.</p><div><hr></div><h2>A Practical Step to Take This Week</h2><p>Here&#8217;s one actionable tip: <strong>pick one AI initiative in your organization and force it through a simple filter.</strong></p><p>Ask:</p><ol><li><p>Does it tie to a clear business outcome?</p></li><li><p>Is our data foundation strong enough to support it?</p></li><li><p>Do our people know how to use and extend it?</p></li><li><p>Can we measure ROI in the next 90 days?</p></li></ol><p>If the answer is &#8220;no&#8221; to more than one, pause and fix the fundamentals before moving forward. You&#8217;ll save resources, but more importantly, you&#8217;ll build credibility when AI projects <em>do</em> succeed.</p><div><hr></div><h2>Your Turn</h2><p>Now, I want to hear from you:</p><ul><li><p>Which of these four mandates feels the hardest in your organization right now and why?</p></li><li><p>Have you seen a project die because leadership treated AI as tech instead of strategy?</p></li><li><p>Or, on the flip side, have you seen a leader set a clear mandate that unlocked real results?</p></li></ul><p>Drop your thoughts in the comments I read every one. The best insights often come from readers like you, not the article itself.</p><p>And remember: <strong>AI isn&#8217;t the strategy. Leadership is.</strong></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://marcosnewsletter.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Marco, Future&#8209;Ops Insights is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[The Death of Junior Hours: Why Consulting Needs an AI Engine Now]]></title><description><![CDATA[Fixed-fee projects are bleeding margins. Here&#8217;s how firms are replacing &#8220;hours&#8221; with an AI department that never sleeps.]]></description><link>https://marcosnewsletter.substack.com/p/the-death-of-junior-hours-why-consulting</link><guid isPermaLink="false">https://marcosnewsletter.substack.com/p/the-death-of-junior-hours-why-consulting</guid><dc:creator><![CDATA[Marco]]></dc:creator><pubDate>Wed, 10 Sep 2025 06:08:22 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!oIPW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ece5e96-dcf4-4caf-9022-93a75a907518_974x687.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Most consulting firms I know are running on a fragile equation: <strong>fixed-fee contracts + armies of junior hours + inconsistent first drafts</strong>. It&#8217;s a formula that worked in the 2000s, when clients tolerated bloated decks and longer timelines. Today, it&#8217;s a slow bleed. Procurement is tougher, margins are thinner, and clients are asking why they should pay for &#8220;training wheels&#8221; time when generative AI can deliver baseline work in minutes.</p><p>And here&#8217;s the uncomfortable truth: in this new equation, <strong>junior hours are worth less than ever, but expectations for speed and quality are higher than ever</strong>.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://marcosnewsletter.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://marcosnewsletter.substack.com/subscribe?"><span>Subscribe now</span></a></p><div><hr></div><h2>The hidden pain every firm is feeling</h2><p>You know the symptoms. A partner signs a fixed-fee deal with a client who&#8217;s already benchmarked you against three other firms. The scope is ambitious, but your slide factory is overloaded. First drafts take weeks. Quality is inconsistent. Teams quietly pad hours just to hit deadlines.</p><p>Margins erode invisibly until the final tally: a project that looked profitable on paper ends up barely covering costs.</p><p>Meanwhile, junior analysts once the backbone of leverage are under fire. Clients push back: <em>&#8220;Why am I paying for four associates to churn slides when I know AI can draft the same content?&#8221;</em> Some procurement teams now refuse to pay for more than a token number of junior hours.</p><p>This leaves firms trapped: you can&#8217;t deliver without them, but you can&#8217;t justify their cost.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!oIPW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ece5e96-dcf4-4caf-9022-93a75a907518_974x687.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!oIPW!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ece5e96-dcf4-4caf-9022-93a75a907518_974x687.png 424w, https://substackcdn.com/image/fetch/$s_!oIPW!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ece5e96-dcf4-4caf-9022-93a75a907518_974x687.png 848w, https://substackcdn.com/image/fetch/$s_!oIPW!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ece5e96-dcf4-4caf-9022-93a75a907518_974x687.png 1272w, https://substackcdn.com/image/fetch/$s_!oIPW!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ece5e96-dcf4-4caf-9022-93a75a907518_974x687.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!oIPW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ece5e96-dcf4-4caf-9022-93a75a907518_974x687.png" width="974" height="687" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1ece5e96-dcf4-4caf-9022-93a75a907518_974x687.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:687,&quot;width&quot;:974,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1374709,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://marcosnewsletter.substack.com/i/172095315?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ece5e96-dcf4-4caf-9022-93a75a907518_974x687.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!oIPW!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ece5e96-dcf4-4caf-9022-93a75a907518_974x687.png 424w, https://substackcdn.com/image/fetch/$s_!oIPW!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ece5e96-dcf4-4caf-9022-93a75a907518_974x687.png 848w, https://substackcdn.com/image/fetch/$s_!oIPW!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ece5e96-dcf4-4caf-9022-93a75a907518_974x687.png 1272w, https://substackcdn.com/image/fetch/$s_!oIPW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1ece5e96-dcf4-4caf-9022-93a75a907518_974x687.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><div><hr></div><h2>You&#8217;re not alone this is systemic</h2><p>This isn&#8217;t just a boutique-firm issue. Look at what&#8217;s happening around the industry:</p><ul><li><p><strong>Klarna</strong> rolled out an AI assistant that now handles <strong>66% of customer chats the equivalent of 700 full-time agents</strong> with higher satisfaction scores than humans.</p></li><li><p><strong>GitHub Copilot</strong>: developers completed coding tasks <strong>55% faster</strong> with AI, in controlled tests.</p></li><li><p><strong>McKinsey&#8217;s 2024 AI report</strong>: 65% of companies already use GenAI regularly, with early adopters reporting measurable cost savings and productivity gains.</p></li></ul><p>The message is clear: AI isn&#8217;t a gadget anymore it&#8217;s a <strong>production engine</strong>. If Klarna can run a customer service department at one-third the cost, why should a CFO keep paying for &#8220;junior slides&#8221;?</p><div><hr></div><h2>The wrong way to use AI in consulting</h2><p>Most firms are still tinkering. A few analysts experiment with ChatGPT on the side. Someone builds a copilot to draft an RFP response. A practice lead tries an off-the-shelf summarizer. These are point solutions useful, but fragile.</p><p>The problem: <strong>AI framed as a team helper, not as a governed profit center</strong>. Without rules, metrics, or pricing models, AI remains a toy. Procurement notices. Clients notice. Eventually, so will your P&amp;L.</p><div><hr></div><h2>A different path: the AI Engagement Engine</h2><p>Imagine flipping the frame. Instead of &#8220;analysts using AI,&#8221; think of <strong>a digital operations layer inside your firm</strong>.</p><p>This layer:</p><ul><li><p>Takes structured briefs and turns them into deliverables competitive teardowns, cost baselines, sensitivity analyses, storyboards.</p></li><li><p>Runs on <strong>A-SLAs</strong> (AI Service Level Agreements) that define latency, confidence thresholds, and escalation rules.</p></li><li><p>Produces transparent metrics: error rates, adherence to templates, factual grounding.</p></li><li><p>Escalates only the tough edge cases to human consultants.</p></li></ul><p>The result: you replace the weakest link unscalable junior hours with a <strong>machine-assisted baseline that&#8217;s faster, cheaper, and defensible to clients</strong>.</p><p>This isn&#8217;t sci-fi. Firms can stand up a pilot in 60 days. Call it an <strong>AI Engagement Engine</strong>: a 24/7 department of micro-services with governance baked in.</p><div><hr></div><h2>A framework to make it real</h2><p>I call it <strong>AMPP</strong> Automate, Manage, Prove, Price.</p><ol><li><p><strong>Automate</strong>: Identify 8&#8211;12 repeatable micro-services (RFP drafts, spend analysis, teardown slides). Define canonical inputs and &#8220;Definition of Done&#8221; outputs.</p></li><li><p><strong>Manage</strong>: Build a RACI around who leads each task (AI-led, human-led, or dual). Set A-SLAs, guardrails for sensitive data, and a &#8220;kill switch&#8221; for rollback.</p></li><li><p><strong>Prove</strong>: Run the engine against 50 historical cases. Measure factuality, adherence to templates, groundedness. Track correction rates.</p></li><li><p><strong>Price</strong>: Shift from hourly billing to <strong>outcome pricing</strong> a floor fee plus a share of automation savings or compressed delivery time.</p></li></ol><p>The pilot canvas can be small: one client, two micro-services, 200 indexed documents, a lead partner plus a QA reviewer. The goal? &lt;30% time to first draft, &lt;10% unnecessary escalations.</p><div><hr></div><h2>Why this matters now</h2><p>Margins are being squeezed harder than ever. Procurement teams are benchmarking everything. EU regulation on AI is tightening. And the truth is brutal: <strong>the firms that turn AI into an operating engine today will dictate tomorrow&#8217;s prices and timelines</strong>.</p><p>Everyone else will be stuck defending hours that clients no longer believe in.</p><div><hr></div><h2>A practical tip you can use this week</h2><p>Don&#8217;t start with a grand strategy. <strong>Pick one repeatable service</strong> that every associate in your practice already dreads say, drafting the baseline of an RFP.</p><ul><li><p>Collect 50 past examples.</p></li><li><p>Feed them into a private RAG index with your firm&#8217;s style guides and templates.</p></li><li><p>Define one simple SLA: if model confidence &lt;0.25, escalate to human review.</p></li><li><p>Time the difference between AI draft vs. manual draft.</p></li></ul><p>That&#8217;s your first proof point. From there, you can scale.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://marcosnewsletter.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Marco, Future&#8209;Ops Insights is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div><hr></div><h2>Your turn</h2><p>I&#8217;m curious:</p><ul><li><p>Do you see your firm still clinging to the &#8220;junior hours&#8221; model?</p></li><li><p>If procurement slashed junior billing tomorrow, how exposed would your margins be?</p></li><li><p>And what&#8217;s the one micro-service you&#8217;d automate first if you had a governed AI engine running in the background?</p></li></ul><p>Drop a comment, I&#8217;d love to hear how your teams are wrestling with this.</p>]]></content:encoded></item><item><title><![CDATA[Archaeology Meets AI and the Ops Lesson on Clarity in Chaos]]></title><description><![CDATA[If AI can uncover lost cities in the jungle, what excuse do we have for not seeing the patterns in our own operations?]]></description><link>https://marcosnewsletter.substack.com/p/archaeology-meets-ai-and-the-ops</link><guid isPermaLink="false">https://marcosnewsletter.substack.com/p/archaeology-meets-ai-and-the-ops</guid><dc:creator><![CDATA[Marco]]></dc:creator><pubDate>Mon, 08 Sep 2025 06:08:22 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!23yn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F841cf7a3-63de-4942-9a64-2e6a9731fc87_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Let&#8217;s begin with something bright and alive: imagine hundreds of digital explorers mostly strangers scattered around the globe, huddled over satellite maps, chat-AI outputs, cluttered desktop screens, all united by a curious itch: <em>what&#8217;s hiding beneath the Amazon canopy?</em> That&#8217;s our starting point: a challenge that is at once technical, creative, and deeply human. It&#8217;s the kind of story that makes you lean in, not because you have to, but because you <em>want</em> to.</p><h2>When Data Feels Like Dirt Under Your Fingernails</h2><p>Here&#8217;s the real pain we often shrug off: wading through endless, complex datasets. Imagine being handed LIDAR scans, multispectral satellite images, temperature readings, colonial-era survey papers, and oral histories from Indigenous communities and told to surface something meaningful. It&#8217;s dense. It&#8217;s messy. It&#8217;s overwhelming.</p><p>You&#8217;ve felt that, right? Tap out "data fatigue" when sorting through CSVs. Or freeze mid-scroll because your visualization isn&#8217;t communicating anything more than your confusion. That moment where you think maybe this mountain of noise will never turn into insight. The <strong>OpenAI to Z community</strong> felt that, too. And they didn&#8217;t flinch.</p><h2>We Are Not Alone: Thousands of Us Facing the Same Wilderness</h2><p>Here&#8217;s what matters: you&#8217;re not the only one staring into the void. Over <strong>8,000 Kagglers</strong> joined the challenge, each wrestling with the same impossibly rich, messy material, each hungering for a breakthrough in the noise. More than <strong>200 submissions</strong>, each a unique lens on the same problem, emerged ranging from predictive maps of potential archaeological hotspots to entirely novel discovery methods born of curiosity and experimentation (<a href="https://www.kaggle.com/c/openai-to-z-challenge?utm_source=chatgpt.com">Kaggle</a>).</p><p>That&#8217;s not a static leaderboard it&#8217;s a forest of approaches, blossoming simultaneously. It&#8217;s hope in iteration. It&#8217;s proof that the moment when you think you&#8217;re alone in your overwhelm, dozens, hundreds or thousands are right there with you.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!23yn!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F841cf7a3-63de-4942-9a64-2e6a9731fc87_1024x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!23yn!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F841cf7a3-63de-4942-9a64-2e6a9731fc87_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!23yn!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F841cf7a3-63de-4942-9a64-2e6a9731fc87_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!23yn!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F841cf7a3-63de-4942-9a64-2e6a9731fc87_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!23yn!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F841cf7a3-63de-4942-9a64-2e6a9731fc87_1024x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!23yn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F841cf7a3-63de-4942-9a64-2e6a9731fc87_1024x1024.png" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/841cf7a3-63de-4942-9a64-2e6a9731fc87_1024x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1800455,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://marcosnewsletter.substack.com/i/172858009?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F841cf7a3-63de-4942-9a64-2e6a9731fc87_1024x1024.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!23yn!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F841cf7a3-63de-4942-9a64-2e6a9731fc87_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!23yn!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F841cf7a3-63de-4942-9a64-2e6a9731fc87_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!23yn!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F841cf7a3-63de-4942-9a64-2e6a9731fc87_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!23yn!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F841cf7a3-63de-4942-9a64-2e6a9731fc87_1024x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2>From Frustration to Flight: The Turning Point</h2><p>Then there&#8217;s the winning team, &#8220;Black Bean,&#8221; who turned that load of raw data and noise into <strong>67 candidate patches</strong>, each about a square mile, identified as promising sites for future archaeological exploration (<a href="https://www.nationalgeographic.com/science/article/amazon-openai-competition-archaeology-machine-learning-artificial-intelligence?utm_source=chatgpt.com">National Geographic</a>).</p><p>They didn&#8217;t just lean on brute-force machine learning. Instead they:</p><ul><li><p>Fused <strong>public LiDAR datasets, Google Earth Engine satellite imagery, NASA elevation data</strong>.</p></li><li><p>Leveraged <strong>GPT-4o</strong> to recognize patterns in known sites and extrapolate to new, unexplored areas especially clustered near waterways, which makes intuitive sense for ancient civilizations (<a href="https://www.nationalgeographic.com/science/article/amazon-openai-competition-archaeology-machine-learning-artificial-intelligence?utm_source=chatgpt.com">National Geographic</a>).</p></li><li><p>Built something that&#8217;s <strong>reproducible</strong> so archaeologists can actually follow their steps.</p></li></ul><p>That is elegance rising from chaos. It&#8217;s the oxygen of clarity breathing through overwhelm. And it&#8217;s organic rooted in reason and grounded in common sense, even as it leveraged bleeding-edge AI.</p><h2>What This Means for You (And Me)</h2><p>What I take away and what might help you is that the impossibly complex problems we face are also opportunities for collective ingenuity. You don&#8217;t have to single-handedly solve it. You can start by:</p><ul><li><p>Embracing complexity instead of shrinking from it.</p></li><li><p>Mining adjacent disciplines or data sources you wouldn&#8217;t normally use.</p></li><li><p>Letting pattern-recognition tools do heavy lifting but never outsourcing your insight.</p></li><li><p>Designing for reproducibility, not just results.</p></li></ul><p>The hackathon wasn&#8217;t about perfect accuracy. It was about creative portfolios multiple ways forward. That&#8217;s your template for experimentation in your daily challenges.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://marcosnewsletter.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://marcosnewsletter.substack.com/subscribe?"><span>Subscribe now</span></a></p><h2>A Tiny Action Tonight That Changes Everything</h2><p>Here&#8217;s a small, immediate move you can make:</p><ol><li><p>pick one project or problem that&#8217;s stalled because of data overwhelm.</p></li><li><p>Ask yourself: <strong>what is </strong><em><strong>one</strong></em><strong> unconventional data source or approach you&#8217;ve dismissed but could actually unlock new perspective?</strong></p></li><li><p>Then spend just <strong>20 minutes</strong> exploring it. Sketch a map of "what might this show me&#8230;?" Use a search engine, a ChatGPT query, a colleague&#8217;s suggestion you&#8217;re not committing, just poking possibilities.</p></li></ol><p>That little nudge let curiosity win, not perfection is enough to spark clarity.</p><div><hr></div><p>So here we are again, at an end that feels like a new beginning. You&#8217;re still overwhelmed, yes, but now you have company, inspiration, and a tiny experiment waiting.</p><p><strong>What unconventional angle might be lying dormant in your work, if only you gave it 20 minutes of curiosity</strong>? I&#8217;d love to hear what sparks for you drop a comment and let&#8217;s ignite the conversation: <strong>what buried connection are </strong><em><strong>you</strong></em><strong> ready to bring to light?</strong></p><div><hr></div><h1>Full Detailed Report on + Executive Summary at the end:</h1><h1>Embracing Complexity and AI in Operations for Professional Services</h1><p><strong>Introduction:</strong><br>In today&#8217;s tech-driven landscape, operations teams especially in professional services face both unprecedented complexity and opportunity. A recent internal <strong>AI hackathon</strong> at a tech firm demonstrated how embracing new AI tools can streamline workflows and spark innovation. However, the hackathon is just a starting point. The real focus is how <strong>operations</strong> in tech and professional service organizations can strategically leverage AI over the next three years. This report examines four key principles for operational excellence in an AI-enabled era, combining strategic vision with practical action. Each principle is grounded in real cases and current practices, showing that these ideas are <strong>feasible</strong> and can lead to concrete improvements rather than being merely theoretical.</p>
      <p>
          <a href="https://marcosnewsletter.substack.com/p/archaeology-meets-ai-and-the-ops">
              Read more
          </a>
      </p>
   ]]></content:encoded></item><item><title><![CDATA[Why 80% of AI projects fail (and how to avoid it)]]></title><description><![CDATA[Most AI initiatives collapse before launch not because of tech, but because of unclear outcomes.]]></description><link>https://marcosnewsletter.substack.com/p/why-80-of-ai-projects-fail-and-how</link><guid isPermaLink="false">https://marcosnewsletter.substack.com/p/why-80-of-ai-projects-fail-and-how</guid><dc:creator><![CDATA[Marco]]></dc:creator><pubDate>Fri, 05 Sep 2025 06:08:26 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/4b79e6af-df2a-4728-87ff-5ea162970f9d_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>It usually starts with excitement. A new budget cycle, a fresh initiative, the CEO inspired by the latest headline about how &#8220;AI is transforming every industry.&#8221; A team gets the green light for an AI project, the consultants are called in, and the mood is optimistic.</p><p>But underneath the enthusiasm, there&#8217;s often something unspoken: the expectation that AI will act like a magic pill. A quick fix for old inefficiencies. A way to leapfrog years of messy operations and finally land in the future.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://marcosnewsletter.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Marco&#8217;s Substack! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>That&#8217;s the myth. And it&#8217;s a dangerous one.</p><div><hr></div><h2>When Expectations Collide with Reality</h2><p>If you&#8217;ve ever been inside one of these projects, you know the story. The kickoff workshop feels full of possibility everyone talking about automation, intelligence, cost savings. But then comes the first real roadblock.</p><ul><li><p>The data isn&#8217;t clean enough.</p></li><li><p>The workflows aren&#8217;t standardized.</p></li><li><p>Nobody can agree on the actual outcome the project should deliver.</p></li></ul><p>What happens next is almost predictable. Meetings drag on. The vendor insists on more custom work. The &#8220;pilot&#8221; never fully launches. And in the end, the project gets shelved quietly, labeled as &#8220;interesting exploration&#8221; rather than the transformation it was supposed to be.</p><p>People rarely admit it, but the real frustration is this: <strong>AI wasn&#8217;t a cure for broken processes. It just made the weaknesses more visible.</strong></p><div><hr></div><h2>You&#8217;re Not Alone: The Data Proves It</h2><p>This isn&#8217;t just anecdotal. Gartner has found that <strong>80% of AI projects never make it past the pilot phase</strong>. Forrester has reported that many initiatives stall because they lack clear business outcomes from the start.</p><p>And when you look closer, the failures don&#8217;t come from the algorithms themselves. They come from human factors: vague goals, unrealistic expectations, and the hope that technology can &#8220;cover&#8221; structural gaps in how work actually happens.</p><p>If you&#8217;ve felt the sting of an AI project that didn&#8217;t live up to the hype, you&#8217;re far from alone. In fact, you&#8217;re in the majority.</p><div><hr></div><h2>The Quick Fix vs. The Long Game</h2><p>The trap is easy to see in hindsight. Companies treat AI like a <strong>patch</strong> a way to fix inefficiencies without addressing the underlying causes.</p><ul><li><p>Need better customer service? Let&#8217;s put AI on top of the chaos.</p></li><li><p>Struggling with supply chain delays? Let&#8217;s add AI dashboards and hope they solve it.</p></li><li><p>Sales data all over the place? Maybe an AI assistant can make sense of it.</p></li></ul><p>But AI isn&#8217;t duct tape. It doesn&#8217;t stick neatly over structural problems. It&#8217;s more like a magnifying glass: it amplifies what&#8217;s already there. If the foundation is shaky, the cracks widen.</p><p>The organizations that actually win with AI approach it differently. They don&#8217;t see it as a shortcut. They see it as a <strong>strategic investment</strong> that only works when paired with clear, measurable outcomes.</p><div><hr></div><h2>What the Winners Do Differently</h2><p>Look at the companies that have made AI more than a buzzword. They start not with &#8220;let&#8217;s add AI,&#8221; but with <strong>&#8220;what outcome do we need to achieve?&#8221;</strong></p><ul><li><p>Reduce manual work so teams get back 10 hours a week.</p></li><li><p>Shorten customer onboarding time by 30%.</p></li><li><p>Increase forecast accuracy by 15%.</p></li></ul><p>Once the outcomes are clear, AI becomes a tool to reach them not the other way around. And suddenly, the hype turns into something more sustainable: progress you can measure.</p><p>The best part? This shift in mindset doesn&#8217;t just make AI more effective it makes teams more aligned. People stop chasing shiny tools and start focusing on results that matter.</p><div><hr></div><h2>A Small, Practical Step for Right Now</h2><p>Here&#8217;s something you can try immediately. Next time an AI proposal crosses your desk, ask one simple question:</p><p><strong>&#8220;What&#8217;s the outcome we&#8217;re committing to, even if we didn&#8217;t use AI?&#8221;</strong></p><p>If the answer is fuzzy, pause. That&#8217;s your signal the project is based on hope, not strategy. If the answer is clear, then AI has a real chance to add value because it&#8217;s tied to something concrete.</p><div><hr></div><h1>Let&#8217;s Open This Up</h1><p>I&#8217;ll leave you with this: the myth of &#8220;AI will fix everything&#8221; isn&#8217;t going away. The headlines will keep coming, the pressure to &#8220;do something with AI&#8221; will stay. But the real opportunity belongs to the people who resist the myth, cut through the noise, and focus on outcomes first.</p><p>Now I want to hear from you:</p><ul><li><p>Have you ever been in an AI project that fell flat what went wrong?</p></li><li><p>Do you think most organizations are chasing AI as a patch, or investing in it as a strategy?</p></li><li><p>What&#8217;s the clearest, most outcome-driven AI initiative you&#8217;ve seen?</p></li></ul><p>Drop your thoughts in the comments I&#8217;m curious to hear your stories and perspectives.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://marcosnewsletter.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Marco&#8217;s Substack! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Professional Services as We Know It Is Dead - Here’s What Comes Next]]></title><description><![CDATA[The AI wave won&#8217;t destroy Professional Services, it will expose who&#8217;s ready to evolve and who gets left behind.]]></description><link>https://marcosnewsletter.substack.com/p/professional-services-as-we-know</link><guid isPermaLink="false">https://marcosnewsletter.substack.com/p/professional-services-as-we-know</guid><dc:creator><![CDATA[Marco]]></dc:creator><pubDate>Wed, 03 Sep 2025 06:08:32 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/d14f1a67-675b-4d5d-9b3b-656dabbaa0f8_1024x1536.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<h1>The Dawn of a Brighter Professional Services Era</h1><p>I hope this finds you curious and cautiously optimistic. Let&#8217;s talk about a shift that feels both unsettling and exciting: what happens when AI starts doing nearly half of what you'd once think was purely human work?</p><h2>The Hidden Load You Carry</h2><p>Imagine your team at the start of the week: juggling scoping documents, detailed delivery spreadsheets, project kickoff decks, status&#8239;updates, invoice checks, repeatable configurations, quality reviews <strong>all the small tasks</strong> that devour time and life. You feel like you&#8217;re firefighting, not strategizing. Maybe you catch yourself muttering, &#8220;If I could spend more time thinking, not just doing&#8230;&#8221;.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://marcosnewsletter.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Marco, Future&#8209;Ops Insights is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>That&#8217;s the <strong>real pain</strong>. When nearly 40&#8211;50% of operational, repeatable tasks especially language-based and documentation-heavy work are swallowed by AI or ready to be, it doesn&#8217;t just shift who does the work. It rattles your sense of purpose in the professional services world.</p><h2>You're Far From Alone</h2><p>You&#8217;re not the only one staring at this transformation. Accenture research estimates that around <strong>40% of working hours</strong> especially in tasks that rely on language can be impacted by LLMs, with about <strong>65%</strong> of those hours ripe for automation or augmentation (<a href="https://www.accenture.com/content/dam/accenture/final/accenture-com/document/Accenture-A-New-Era-of-Generative-AI-for-Everyone.pdf?utm_source=chatgpt.com">Accenture</a>).</p><p>And this isn&#8217;t some distant, hypothetical future. Accenture reports that <strong>74% of organizations</strong> investing in generative AI and automation are meeting or exceeding expectations and these &#8220;reinvention&#8209;ready&#8221; firms are already achieving <strong>2.4&#215; productivity</strong> improvements and <strong>2.5&#215; revenue growth</strong> (<a href="https://newsroom.accenture.com/news/2024/new-accenture-research-finds-that-companies-with-ai-led-processes-outperform-peers?utm_source=chatgpt.com">Accenture Newsroom</a>).</p><p>TSIA points at the move toward AI&#8209;powered, subscription&#8209;oriented, outcome&#8209;centric models what they call PS 2.0 as the way forward. It&#8217;s not just about automation, but reshaping delivery models to continuously drive measurable outcomes (<a href="https://www.tsia.com/research/professional-services-2-0?utm_source=chatgpt.com">TSIA</a>).</p><h2>When Systems Do the Routine, You Can Lead</h2><p>The light at the end of the tunnel: AI doesn&#8217;t erase need for human insight it frees us to focus on <strong>strategy, creativity, empathy, interpretation</strong>. It demands new roles, new expertise, and new business models.</p><p>Accenture&#8217;s &#8220;reinvention services&#8221; is a case in point they&#8217;re blending strategy, consulting, tech, operations into one AI&#8209;powered unit. One client example? An AI&#8209;enabled ship that predicts its own maintenance, manages energy, and communicates with ports (<a href="https://www.businessinsider.com/accenture-ai-reinvention-services-earnings-ceo-julie-sweet-2025-6?utm_source=chatgpt.com">Business Insider</a>). That&#8217;s not replacing people; it&#8217;s redefining how people and systems collaborate.</p><p>And this isn&#8217;t theoretical. The World Economic Forum notes that roles like AI/ML specialists, data analysts and transformation strategists are <strong>growing rapidly</strong>, even as more clerical or secretarial duties recede (<a href="https://www.weforum.org/stories/2023/05/jobs-lost-created-ai-gpt/?utm_source=chatgpt.com">World Economic Forum</a>).</p><h2>What&#8217;s the Path Forward In Practice</h2><p>Here&#8217;s where we zoom in. Skipping flashy promises for grounded, purposeful strategy:</p><ol><li><p><strong>Redesign your delivery model</strong>: Shift from &#8220;billable hours&#8221; to &#8220;PS&#8209;as&#8209;a&#8209;service,&#8221; where outcomes are the unit of value (<a href="https://www.tsia.com/research/professional-services-2-0?utm_source=chatgpt.com">TSIA</a>).</p></li><li><p><strong>Conceive new roles</strong>: Intelligence&#8209;curators, AI&#8209;workflow architects, client outcome designers roles that blend domain knowledge with AI&#8209;fluency.</p></li><li><p><strong>Build your AI governance and strategy stack</strong>: Align senior leadership, roadmap, KPIs, training, and feedback loops. McKinsey finds this kind of setup correlates strongly with actual bottom&#8209;line impact (<a href="https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai?utm_source=chatgpt.com">McKinsey &amp; Company</a>).</p></li><li><p><strong>Embed learning and incentives</strong>: Fed by Accenture and McKinsey findings, employees are eager to use AI but they need training, seamless integration, and encouragement, not just mandates (<a href="https://www.mckinsey.com/capabilities/mckinsey-digital/our-insights/superagency-in-the-workplace-empowering-people-to-unlock-ais-full-potential-at-work?utm_source=chatgpt.com">McKinsey &amp; Company</a>, <a href="https://www.accenture.com/us-en/insights/consulting/gen-ai-reinventing-enterprise-models?utm_source=chatgpt.com">Accenture</a>).</p></li></ol><p>This isn&#8217;t about reacting to AI&#8217;s arrival. It&#8217;s a deliberate invitation to <strong>reinvent how we deliver value</strong> in professional services.</p><div><hr></div><h3>Action Happens Now A Practical Kick-Start</h3><p>Here&#8217;s a <strong>concrete move you can take today</strong>: <strong>run a 90-minute &#8220;AI-task mapping session&#8221; with your team.</strong> Grab a whiteboard (virtual or physical). Ask: <em>Which tasks do we repeat most? Which bleed us dry?</em> Then, categorize them:</p><ul><li><p>Automatable</p></li><li><p>Augmentable</p></li><li><p>High-value, human-only work</p></li></ul><p>Within an hour and a half, you&#8217;ll not only illuminate where AI could rescue bandwidth, but also discover where your people&#8217;s strategic thinking should go next. That simple session often unlocks the jump to PS 2.0.</p><p>You&#8217;re not just surviving this transition you can steer it.</p><p>I&#8217;d love to hear from you: <strong>In your work today, where do you already feel the drag of repetitive tasks? What would you do if 50% of that workload were suddenly handled for you?</strong> Share your examples or wildest dreams in the comments. Let&#8217;s unpack what &#8220;reinvention&#8221; could look like, together.</p><p>Here&#8217;s to building smarter, more human-centered professional services.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://marcosnewsletter.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Marco, Future&#8209;Ops Insights is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item></channel></rss>