<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Context Engineering]]></title><description><![CDATA[What actually works when deploying AI at enterprise scale. One observation from the field, every week.]]></description><link>https://matthewkruczek.substack.com</link><image><url>https://substackcdn.com/image/fetch/$s_!VXWr!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec4aca9e-0c9d-493e-b1ff-064c3a072af8_736x736.png</url><title>Context Engineering</title><link>https://matthewkruczek.substack.com</link></image><generator>Substack</generator><lastBuildDate>Mon, 06 Apr 2026 23:04:55 GMT</lastBuildDate><atom:link href="https://matthewkruczek.substack.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Matthew kruczem]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[matthewkruczek@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[matthewkruczek@substack.com]]></itunes:email><itunes:name><![CDATA[Matthew Kruczek]]></itunes:name></itunes:owner><itunes:author><![CDATA[Matthew Kruczek]]></itunes:author><googleplay:owner><![CDATA[matthewkruczek@substack.com]]></googleplay:owner><googleplay:email><![CDATA[matthewkruczek@substack.com]]></googleplay:email><googleplay:author><![CDATA[Matthew Kruczek]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[The AI Readiness Scorecard: Eight Dimensions That Determine Whether You Scale or Stall]]></title><description><![CDATA[If you only have a minute, here&#8217;s what you need to know.]]></description><link>https://matthewkruczek.substack.com/p/the-ai-readiness-scorecard-eight</link><guid isPermaLink="false">https://matthewkruczek.substack.com/p/the-ai-readiness-scorecard-eight</guid><dc:creator><![CDATA[Matthew Kruczek]]></dc:creator><pubDate>Mon, 06 Apr 2026 16:57:51 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!VXWr!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec4aca9e-0c9d-493e-b1ff-064c3a072af8_736x736.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>If you only have a minute, here&#8217;s what you need to know.</em></p><ul><li><p>Most organizations buying AI tools have never assessed whether their organization can absorb them. Cisco&#8217;s study of 8,000 companies found only 13% are actually ready to deploy AI at scale. The other 87% are spending money on capability they can&#8217;t operationalize.</p></li><li><p>AI readiness isn&#8217;t one thing. It&#8217;s eight: Leadership Commitment, Strategic Alignment, Data Readiness, Technology Infrastructure, Talent &amp; Skills, Process Maturity, Culture &amp; Change Readiness, and Governance &amp; Ethics. Weakness in any single dimension blocks the others.</p></li><li><p>Two of those dimensions, Leadership and Strategic Alignment, function as multipliers. Organizations that score high on those two extract more value from every dollar invested in the other six. Organizations that score low waste money regardless of how advanced their technology is.</p></li><li><p>Industry benchmarks exist: financial services typically scores 2.8-3.5 out of 5.0, healthcare 2.0-2.8, manufacturing 2.2-3.0. Any dimension scoring below 3.0 acts as a blocker, no matter how strong the rest of the scorecard looks.</p></li><li><p>This article gives you enough to sketch your own radar chart and find your weak spots. The rest of this series will show you what to do about each one.</p></li></ul><div><hr></div><p>I&#8217;ve <a href="https://matthewkruczek.ai/blog/ai-adoption-paradox">written before</a> about why AI pilots fail, and why only 10% of success comes from the technology itself. Most executives I talk to already know their AI initiatives are underperforming. What they don&#8217;t know is <em>where exactly</em> their organization is breaking down and which problems to fix first.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://matthewkruczek.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Context Engineering! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>That&#8217;s a diagnosis problem. And almost nobody is doing it.</p><h2><strong>The purchase order that replaces strategy</strong></h2><p>Here&#8217;s the pattern I see repeatedly. A leadership team reads the headlines, feels the competitive pressure, and makes a purchasing decision. Copilot licenses for 5,000 employees. An Azure OpenAI deployment. A handful of pilot projects with aggressive timelines.</p><p>Six months later, Copilot adoption is at 15%. The pilots produced impressive demos that nobody scaled. The Azure deployment is running one use case that could have been built with a Python script. The organization spent millions on AI capability and has almost nothing to show for it.</p><p>The problem isn&#8217;t the tools. The problem is that nobody asked whether the organization was ready for them.</p><p>Cisco&#8217;s 2025 AI Readiness Index surveyed 8,000 organizations globally. Only 13% qualified as &#8220;Pacesetters,&#8221; companies actually converting AI investment into production value. Those Pacesetters were 4x more likely to move AI from pilot to production than everyone else. The gap wasn&#8217;t technology budgets or model selection. It was organizational readiness across multiple dimensions simultaneously.</p><p>The organizations that succeed assessed their readiness before they started buying. The ones that struggle skipped diagnosis and went straight to procurement.</p><h2><strong>Eight dimensions, not one</strong></h2><p>AI readiness isn&#8217;t a single score. It&#8217;s a profile across eight distinct dimensions, and weakness in any one of them can block progress in all the others.</p><p>I&#8217;ve been refining this framework across multiple enterprise engagements. The eight dimensions that consistently determine whether an organization scales or stalls:</p><h3><strong>1. Leadership Commitment</strong></h3><p>This is the dimension most organizations overestimate. Having a CEO who talks about AI in earnings calls is not leadership commitment.</p><p>Leadership commitment means a named executive sponsor with dedicated time and authority over AI initiatives. Not &#8220;the CIO will oversee this alongside their other responsibilities.&#8221; A person whose calendar reflects AI transformation as a primary obligation, not an afterthought. It means an AI steering committee with cross-functional representation, because AI that lives only in IT never reaches the business units where value is created. And it means a dedicated budget line item that isn&#8217;t buried in general IT spend, because initiatives funded through discretionary budget get cut the first time the quarter looks tight.</p><p>Here&#8217;s what a 2.0 looks like: the CEO mentioned AI at the last all-hands, someone in IT was told to &#8220;look into it,&#8221; and there&#8217;s no dedicated budget. A 4.0 looks different: a named sponsor reports to the board quarterly on AI progress, there&#8217;s a cross-functional steering committee that meets biweekly, and AI has its own P&amp;L line that survived two budget cycles.</p><p>Sustained sponsorship correlates with a 68% success rate. Without it, success drops to 11%. More than half of AI initiatives lose their executive sponsor within six months, usually because the sponsor treated it as a ribbon-cutting rather than a multi-year commitment. The next article in this series goes deep on what the right sponsor looks like and why most companies get this wrong.</p><h3><strong>2. Strategic Alignment</strong></h3><p>Most AI roadmaps are technology wish lists. They catalog tools the organization wants to buy rather than business problems AI should solve.</p><p>Strategic alignment means linking every AI initiative to one of the company&#8217;s top three to five business priorities. Not to vague goals like &#8220;explore AI&#8221; or &#8220;increase innovation.&#8221; If your CEO&#8217;s top priority is margin expansion in the North American retail division, your first AI initiative should trace a direct line to that outcome. If it doesn&#8217;t, it won&#8217;t survive the first budget review.</p><p>It also means having a use-case prioritization framework that weighs feasibility against impact against data readiness. Without one, organizations pursue whatever the loudest executive asks for or whatever the vendor demo looked most impressive. That&#8217;s how you end up with 20 teams building 20 disconnected chatbots.</p><p>A 2.0 here looks like a list of 30 AI &#8220;opportunities&#8221; with no ranking criteria and no connection to business outcomes. A 4.0 looks like five prioritized use cases, each mapped to a business metric, with clear owners and quarterly checkpoints. The difference isn&#8217;t sophistication. It&#8217;s discipline.</p><h3><strong>3. Data Readiness</strong></h3><p>This is where most AI projects actually die. Not in model selection, not in deployment, but in the data underneath. 85% of AI projects fail due to poor data quality or lack of relevant data.</p><p>Data readiness covers three things. First, quality: is the data accurate, complete, and current? Most organizations discover their data is far dirtier than they assumed once they try to feed it to a model. Second, accessibility: is the data centralized on a platform where AI services can reach it, or is it trapped in departmental silos, legacy databases, and someone&#8217;s Excel spreadsheet? Third, governance: are permissions set correctly, is sensitive data classified, and do you know who can access what?</p><p>That third one is where Microsoft Copilot deployments have been a wake-up call. Multiple enterprises discovered that rolling out Copilot exposed over-permissioned SharePoint sites, giving employees AI-surfaced access to documents they were never supposed to see. The technology worked perfectly. The data governance underneath it didn&#8217;t. Those organizations spent months on permission audits before the tool could be safely used.</p><p>Among top-performing organizations, 76% have fully centralized their data. Among everyone else, only 19%. That single gap explains more of the performance difference than any technology choice.</p><h3><strong>4. Technology Infrastructure</strong></h3><p>This is the dimension that gets the most attention and probably matters the least, at least in isolation.</p><p>Cloud capacity, a standardized AI/ML platform, an integration layer that connects AI services to existing systems, API management, compute scaling. These are necessary. They are nowhere near sufficient. The platforms are mature. Azure AI Foundry, Copilot Studio, Microsoft Fabric, AWS Bedrock, Google Vertex, and their equivalents provide more than enough technical foundation for any enterprise AI initiative shipping today.</p><p>Technology accounts for roughly 10% of what determines AI success. Yet it&#8217;s where most organizations start and where most of the budget goes. This isn&#8217;t an argument against investing in infrastructure. It&#8217;s an argument against investing in infrastructure <em>first</em>, before you&#8217;ve addressed the seven dimensions that determine whether anyone uses what you build.</p><p>A 2.0 is running AI experiments on individual developers&#8217; laptops with no shared platform. A 4.0 is a standardized enterprise platform with self-service provisioning, monitoring, cost management, and integration connectors to your core systems. Most organizations are somewhere around 3.0 here already, which is fine. The bottleneck is almost never the platform.</p><h3><strong>5. Talent &amp; Skills</strong></h3><p>A one-day workshop on prompt engineering changes nothing. I&#8217;ve watched organizations spend six figures on AI training programs that produce certificates and zero behavior change.</p><p>Talent and skills covers three layers. The first is AI literacy across the organization: does the average employee understand what AI can and can&#8217;t do, and do they feel confident experimenting with it in their daily work? The second is practitioner depth: do you have dedicated AI/ML engineers, data scientists, and platform engineers who can build and maintain AI systems? The third is cross-functional fluency: can business analysts, product managers, and domain experts articulate problems in ways that translate to AI solutions?</p><p>Role-based training matters. An executive needs to understand AI&#8217;s strategic implications and governance requirements. A middle manager needs to know how to redesign their team&#8217;s workflows around AI capabilities. A software engineer needs hands-on experience with the tools and frameworks. A finance analyst needs practical training on AI-assisted analysis within their actual tools. One curriculum for all four audiences will fail all four.</p><p>Among top-performing organizations, 75% report AI proficiency across their staff. Among everyone else, 16%. That&#8217;s the widest gap of any dimension in the data, and it&#8217;s not a gap you close with a lunch-and-learn series.</p><h3><strong>6. Process Maturity</strong></h3><p>AI can&#8217;t optimize processes that aren&#8217;t documented, standardized, or measured. This is the dimension that catches organizations off guard.</p><p>Here&#8217;s a test. Pick any core business process in your organization, something like &#8220;how we approve a new vendor&#8221; or &#8220;how we onboard a new customer.&#8221; Now ask three people in three different offices to describe the steps. If you get three different answers, your process maturity score is below 3.0, and AI will amplify the inconsistency rather than fix it.</p><p>The target state is making your implicit operating model explicit and machine-readable. That&#8217;s where AI becomes transformative, when it can read your processes, identify bottlenecks, and suggest or execute improvements. But the starting point is much simpler: are your core business processes documented? Are they standardized across teams and geographies? Are they measured with KPIs that would tell you whether AI is actually improving them?</p><p>If the answer is no, AI deployment will produce anecdotes, not outcomes. You&#8217;ll have impressive demos of AI &#8220;optimizing&#8221; a process that doesn&#8217;t exist in any consistent form, and no way to measure whether the optimization made a difference.</p><h3><strong>7. Culture &amp; Change Readiness</strong></h3><p>I&#8217;ve written about the <a href="https://matthewkruczek.ai/blog/ai-adoption-paradox">AI adoption paradox</a> in detail, including the finding that 31% of employees actively sabotage AI initiatives. Culture isn&#8217;t a soft dimension. It&#8217;s the one that determines whether your other seven investments get used or ignored.</p><p>Culture and change readiness encompasses three things. Psychological safety: do employees feel safe experimenting with AI, including producing bad outputs while learning? Change management: does the organization have a plan for how roles, workflows, and expectations will evolve, or is the implicit message &#8220;figure it out&#8221;? And tolerance for failure: when an AI pilot doesn&#8217;t work, does the organization learn from it or kill the program?</p><p>The organizations that achieve the fastest adoption make it opt-in, not mandated. They create environments where early adopters become visible champions, where success stories spread organically, and where people adopt because they see peers getting better outcomes, not because they received a compliance email. Culture change through pull is more durable than culture change through push. But you have to know whether your culture can support that before you design the rollout.</p><p>A 2.0 looks like active resistance: employees avoiding AI tools, managers vocally skeptical in meetings, no change management plan. A 4.0 looks like organic experimentation: employees sharing tips in Slack channels, managers redesigning team workflows around AI, and a formal change management lead coordinating the transition.</p><h3><strong>8. Governance &amp; Ethics</strong></h3><p>Most organizations frame this as a tradeoff: governance or speed. That&#8217;s a false choice. The answer isn&#8217;t less governance. It&#8217;s governance designed for velocity.</p><p>This dimension covers four areas. Principles: has the organization documented what responsible AI use looks like, in specific terms, not a vague values statement? Policies: is there an acceptable use policy that employees actually know about, covering what data can go into AI tools, what outputs require human review, and what&#8217;s off limits? Process: is there a review mechanism for new AI applications before they go into production, and can it turn around a decision in days rather than months? And identity: as AI agents start acting on behalf of the organization, who are they, what can they access, and who&#8217;s accountable for what they do?</p><p>That last area is evolving fast. Microsoft&#8217;s Entra Agent ID is one example of where this is heading: verified identity for every AI agent, not just every human user. Your governance framework needs to account for non-human actors that can read, write, and transact.</p><p>Here&#8217;s the reality most leaders haven&#8217;t fully absorbed: 68% of employees are already using AI tools without IT approval. Your people aren&#8217;t waiting for your governance framework. They&#8217;re working around it. The question isn&#8217;t whether you need governance. It&#8217;s whether your governance can move fast enough that people actually use the governed path instead of the shadow one.</p><h2><strong>The multiplier effect</strong></h2><p>Not all eight dimensions carry equal weight. Two of them, Leadership Commitment and Strategic Alignment, function as multipliers on everything else.</p><p>This makes intuitive sense. A brilliant data strategy accomplishes nothing if leadership isn&#8217;t committed to funding it through the 6-12 months it takes to operationalize. A world-class AI platform collects dust if there&#8217;s no strategic alignment on which problems to solve with it. Conversely, strong leadership and clear strategic direction amplify every other investment.</p><p>In the scoring model I use, Leadership and Strategic Alignment each carry 1.5x weight. The formula looks like this:</p><blockquote><p><strong>Overall Score = (1.5 x Leadership + 1.5 x Strategic Alignment + Data + Technology + Talent + Process + Culture + Governance) / 9</strong></p><p>Scale: 1.0 to 5.0.</p></blockquote><p>This weighting means an organization scoring 4.0 on leadership and strategy but 3.0 on everything else will outperform an organization scoring 3.0 on leadership and strategy but 4.0 on everything else. The multipliers matter more than the averages.</p><h2><strong>Where industries typically score</strong></h2><p>Here&#8217;s where industries tend to land based on the assessments I&#8217;ve seen and the broader research data:</p><ul><li><p><strong>Financial services: 2.8 to 3.5.</strong> Strongest in governance (3.5-4.5), reflecting decades of regulatory compliance muscle. Weakest in culture (2.0-3.0), where risk aversion that serves them well in banking works against them in AI experimentation.</p></li><li><p><strong>Healthcare: 2.0 to 2.8.</strong> Strongest in governance (2.5-3.5), driven by HIPAA and clinical trial requirements. Weakest in data readiness (1.5-2.5), where fragmented EHR systems and interoperability challenges create a data infrastructure problem that predates AI by decades.</p></li><li><p><strong>Manufacturing: 2.2 to 3.0.</strong> Strongest in process maturity (3.0-4.0), because manufacturing has been documenting and measuring processes since before software existed. Weakest in talent (1.5-2.5), where the AI skills gap intersects with an existing shortage of digital talent on factory floors.</p></li><li><p><strong>Professional services: 2.3 to 3.2.</strong> Strongest in talent (3.0-4.0), because knowledge workers tend to adopt new tools faster. Weakest in technology infrastructure (2.0-3.0), where years of underinvestment in platforms create a foundation gap.</p></li></ul><p>Company size matters too. Small companies (50M&#8722;50<em>M</em>&#8722;200M revenue) typically score 1.8-2.5, constrained primarily by talent. Mid-market companies (200M&#8722;200<em>M</em>&#8722;1B) score 2.3-3.2, with data readiness as the most variable dimension. Enterprise companies ($1B+) score 2.8-3.8 but struggle with organizational complexity and change fatigue.</p><p>These are ranges, not rules. But they give you a starting point for honest self-assessment.</p><h2><strong>The blocking rule</strong></h2><p>Here&#8217;s the finding that makes executives most uncomfortable: any single dimension scoring below 3.0 blocks transformation, regardless of how strong the other dimensions look.</p><p>A financial services company with a 4.5 in governance and a 2.0 in culture will stall. Their compliance framework is excellent, but employees won&#8217;t adopt the tools. A technology company with a 4.5 in talent and a 2.0 in data readiness will stall. Their engineers are eager and capable, but there&#8217;s nothing clean to feed the models.</p><p>This is why holistic assessment matters. Organizations naturally invest in their strengths, the dimensions where they already score well, because progress feels easiest there. But the scorecard makes visible a different truth: the highest-ROI investment is always in your weakest dimension, because that&#8217;s the one blocking everything else.</p><p>The pattern among top performers isn&#8217;t one dominant strength. It&#8217;s the absence of critical weaknesses. They score above threshold on every dimension simultaneously.</p><h2><strong>What to do this week</strong></h2><p>You don&#8217;t need a consulting engagement to start. You need a whiteboard, 90 minutes, and the willingness to be honest.</p><p><strong>Score each dimension.</strong> Pull together a cross-functional group, not just IT, and rate each of the eight dimensions on a 1-5 scale. Use the descriptions above as rough calibration. Where you find disagreement within the room, you&#8217;ve found a gap in organizational awareness, which is itself a finding.</p><p><strong>Plot the radar chart.</strong> Eight axes, one per dimension. The shape tells you more than the average. A spiky chart (high in some areas, low in others) indicates an organization investing unevenly. A uniformly low chart indicates an organization early in the journey. A chart with one or two deep dips below 3.0 shows you exactly what&#8217;s blocking progress.</p><p><strong>Find your blocker.</strong> Look for any dimension below 3.0. That&#8217;s your first priority, not because it&#8217;s the most exciting work, but because nothing else moves until it does. Leadership and Strategic Alignment below 3.0 are the most urgent because they&#8217;re multipliers on everything else.</p><p><strong>Stop buying tools for six weeks.</strong> This is the uncomfortable recommendation. If you haven&#8217;t done this assessment, pause net-new AI tool purchases until you have. Every tool you buy before understanding your readiness profile is a bet that your organization can absorb it. The data says that&#8217;s a bet you&#8217;ll lose 87% of the time.</p><p>This is the first article in <strong>The AI Readiness Playbook</strong> series. The next eight articles will walk through how to close the gaps this scorecard reveals, starting with executive sponsorship and strategic alignment, through data readiness, governance, skills, engineering enablement, and the messy middle of scaling from pilots to production.</p><p>The scorecard tells you where you are. The playbook tells you how to move.</p><div><hr></div><p><em>Matthew Kruczek is Managing Director at EY, leading Microsoft domain initiatives within Digital Engineering. Connect with Matthew on <a href="https://www.linkedin.com/in/matthew-kruczek/">LinkedIn</a> to discuss AI readiness assessment and organizational enablement for your enterprise.</em></p><h2><strong>References</strong></h2><ol><li><p>Cisco. &#8220;AI Readiness Index 2025.&#8221; 8,000 organizations surveyed globally. <a href="https://www.cisco.com/c/m/en_us/solutions/ai/readiness-index.html">cisco.com</a></p></li><li><p>BCG. &#8220;From Potential to Profit: Closing the AI Impact Gap.&#8221; January 2025. <a href="https://www.bcg.com/publications/2025/closing-the-ai-impact-gap">bcg.com</a></p></li><li><p>BCG. &#8220;Enterprise as Code: Operating Model for the AI Era.&#8221; December 2025. <a href="https://www.bcg.com/publications/2025/enterprise-as-code-operating-model-for-ai-era">bcg.com</a></p></li><li><p>Gartner. &#8220;AI Maturity Model and Roadmap Toolkit.&#8221; <a href="https://www.gartner.com/en/chief-information-officer/research/ai-maturity-model-toolkit">gartner.com</a></p></li><li><p>Microsoft. &#8220;Enterprise AI Maturity in Five Steps.&#8221; October 2025. <a href="https://www.microsoft.com/insidetrack/blog/enterprise-ai-maturity-in-five-steps-our-guide-for-it-leaders/">microsoft.com</a></p></li><li><p>Kruczek, M. &#8220;The AI Adoption Paradox.&#8221; <a href="https://matthewkruczek.ai/blog/ai-adoption-paradox">matthewkruczek.ai</a></p></li></ol><div><hr></div><p><em>This is Article 1 of 9 in &#8220;The AI Readiness Playbook&#8221; series, a step-by-step methodology for making your organization AI-ready.</em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://matthewkruczek.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Context Engineering! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Same Prompt, Different Results. Your Agent Harness Is the Multiplier]]></title><description><![CDATA[If you only have a minute, here&#8217;s what you need to know.]]></description><link>https://matthewkruczek.substack.com/p/same-prompt-different-results-your</link><guid isPermaLink="false">https://matthewkruczek.substack.com/p/same-prompt-different-results-your</guid><dc:creator><![CDATA[Matthew Kruczek]]></dc:creator><pubDate>Mon, 30 Mar 2026 17:51:43 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!SIw7!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5770cd2f-d594-4c80-b6c5-6e2f84738118_900x530.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>If you only have a minute, here&#8217;s what you need to know.</em></p><ul><li><p>Most engineers run AI coding agents with default settings. They&#8217;re leaving 60-70% of the tool&#8217;s capability on the table.</p></li><li><p>An <strong>agent harness</strong>, the rules, hooks, skills, memory systems, and plugins wrapping your AI agent, determines output quality more than the model itself. Same model, same prompt, dramatically different results.</p></li><li><p>My Claude Code harness uses 16 lifecycle hooks, 28 reusable skills, 5 plugins, 8 MCP server endpoints (5 aggregated through a single gateway), and a cross-project memory graph. Every layer is designed around one principle: <strong>optimize what enters the context window.</strong></p></li><li><p>Token comparisons show a <strong>3-4x efficiency gain</strong> on research-heavy tasks and a <strong>2x reduction in correction cycles</strong> on implementation tasks, compared to default configuration.</p></li><li><p>The context window is the bottleneck for all agentic coding. You don&#8217;t need my exact setup. But if you&#8217;re running an AI coding agent without governance rules, memory persistence, and context optimization, you&#8217;re paying for a sports car and driving it in first gear.</p></li></ul><div><hr></div><p>I wrote two weeks ago that <a href="https://matthewkruczek.substack.com/p/agent-harnesses-dont-need-more-layers">agent harnesses need fewer layers, not more</a>. That argument still stands. But &#8220;fewer layers&#8221; doesn&#8217;t mean &#8220;no layers.&#8221; It means <em>the right layers</em>.</p><p>Today I want to show you what that looks like in practice. Not a framework diagram. Not a theoretical architecture. My actual Claude Code configuration, the one I use every day to ship production code, monitor my social media, and manage a portfolio of 30+ projects.</p><p>This isn&#8217;t a tutorial. It&#8217;s an argument: <strong>your agent harness is the single highest-leverage investment you can make in AI-augmented engineering</strong>, and most teams haven&#8217;t even started.</p><h2><strong>Why the context window is everything</strong></h2><p>Before I walk through my setup, I need to make the case for why all of this matters. The answer is three words: <strong>the context window.</strong></p><p>Every AI coding agent operates within a fixed context window. That&#8217;s the token budget holding system instructions, your conversation history, tool outputs, and the model&#8217;s own reasoning. When that window fills up, the agent compacts: it summarizes and discards older context to make room. Every compaction loses conversational state. Every lost state means re-explanation, re-exploration, and re-orientation.</p><p>The default Claude Code experience treats the context window as a dumb pipe. Raw build logs? Dump them in. Five hundred lines of <code>git log</code>? Dump them in. The model&#8217;s own correction attempts after generating code that doesn&#8217;t match your standards? All of it goes in. Most sessions, 60-75% of the context window is consumed by information the model doesn&#8217;t need in raw form.</p><p>This is the fundamental problem a harness solves. <strong>Every layer of a production harness exists to optimize what enters the context window.</strong> Front-load the right information, sandbox the noise, and eliminate the correction cycles that waste tokens on work the model should have gotten right the first time.</p><p>The model is powerful. The context window is finite. The harness bridges that gap.</p><h2><strong>What &#8220;default&#8221; actually looks like</strong></h2><p>When you install Claude Code and start a conversation, here&#8217;s what happens:</p><p>The model receives your prompt. It has access to built-in tools: file read, file write, bash execution, web search. It knows nothing about your project conventions, your coding style, your team&#8217;s architectural decisions, or what you worked on yesterday. Every session starts from zero.</p><p>This is like hiring a senior engineer and giving them no onboarding, no documentation, no code review standards, and no access to your team&#8217;s Slack history. They&#8217;re talented. They&#8217;ll produce something. But they&#8217;ll spend half their time asking questions you&#8217;ve already answered and writing code that doesn&#8217;t match your patterns.</p><p>That&#8217;s the default experience. And it&#8217;s what 90% of Claude Code users are running right now.</p><p>Here&#8217;s where their context window goes in a typical 2-hour session:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!SIw7!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5770cd2f-d594-4c80-b6c5-6e2f84738118_900x530.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!SIw7!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5770cd2f-d594-4c80-b6c5-6e2f84738118_900x530.png 424w, https://substackcdn.com/image/fetch/$s_!SIw7!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5770cd2f-d594-4c80-b6c5-6e2f84738118_900x530.png 848w, https://substackcdn.com/image/fetch/$s_!SIw7!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5770cd2f-d594-4c80-b6c5-6e2f84738118_900x530.png 1272w, https://substackcdn.com/image/fetch/$s_!SIw7!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5770cd2f-d594-4c80-b6c5-6e2f84738118_900x530.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!SIw7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5770cd2f-d594-4c80-b6c5-6e2f84738118_900x530.png" width="900" height="530" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5770cd2f-d594-4c80-b6c5-6e2f84738118_900x530.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:530,&quot;width&quot;:900,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:39485,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://matthewkruczek.substack.com/i/192637819?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5770cd2f-d594-4c80-b6c5-6e2f84738118_900x530.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!SIw7!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5770cd2f-d594-4c80-b6c5-6e2f84738118_900x530.png 424w, https://substackcdn.com/image/fetch/$s_!SIw7!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5770cd2f-d594-4c80-b6c5-6e2f84738118_900x530.png 848w, https://substackcdn.com/image/fetch/$s_!SIw7!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5770cd2f-d594-4c80-b6c5-6e2f84738118_900x530.png 1272w, https://substackcdn.com/image/fetch/$s_!SIw7!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5770cd2f-d594-4c80-b6c5-6e2f84738118_900x530.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Less than half the context window is doing real work. The rest is overhead. That&#8217;s the cost of having no harness.</p><h2><strong>The seven layers of a production harness</strong></h2><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2Gl9!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8479e35-c55c-42c7-8b04-02fa31465416_1200x630.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2Gl9!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8479e35-c55c-42c7-8b04-02fa31465416_1200x630.png 424w, https://substackcdn.com/image/fetch/$s_!2Gl9!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8479e35-c55c-42c7-8b04-02fa31465416_1200x630.png 848w, https://substackcdn.com/image/fetch/$s_!2Gl9!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8479e35-c55c-42c7-8b04-02fa31465416_1200x630.png 1272w, https://substackcdn.com/image/fetch/$s_!2Gl9!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8479e35-c55c-42c7-8b04-02fa31465416_1200x630.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2Gl9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8479e35-c55c-42c7-8b04-02fa31465416_1200x630.png" width="1200" height="630" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d8479e35-c55c-42c7-8b04-02fa31465416_1200x630.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:630,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:235843,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://matthewkruczek.substack.com/i/192637819?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8479e35-c55c-42c7-8b04-02fa31465416_1200x630.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!2Gl9!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8479e35-c55c-42c7-8b04-02fa31465416_1200x630.png 424w, https://substackcdn.com/image/fetch/$s_!2Gl9!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8479e35-c55c-42c7-8b04-02fa31465416_1200x630.png 848w, https://substackcdn.com/image/fetch/$s_!2Gl9!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8479e35-c55c-42c7-8b04-02fa31465416_1200x630.png 1272w, https://substackcdn.com/image/fetch/$s_!2Gl9!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd8479e35-c55c-42c7-8b04-02fa31465416_1200x630.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>My harness has seven layers. Each one solves a specific problem that default configuration doesn&#8217;t address, and each one is fundamentally about putting the right tokens into the context window while keeping the wrong ones out.</p><h3><strong>Layer 1: Governance rules (7 enforcement files)</strong></h3><p>The <code>~/.claude/rules/</code> directory contains seven markdown files that load into every conversation as system instructions. They encode:</p><ul><li><p><strong>Coding style.</strong> Immutability patterns for TypeScript and C#. Spread operators, never mutate. Records and <code>with</code> expressions. Init-only properties. This isn&#8217;t a preference; it&#8217;s a requirement enforced before the first line of code is written.</p></li><li><p><strong>Security checklist.</strong> No hardcoded secrets. FluentValidation on all DTOs. EF Core only for SQL (auto-parameterized). JWT validation with <code>ClockSkew=Zero</code>. CSRF protection. Rate limiting on public endpoints. This runs as a pre-commit mental model, not an afterthought.</p></li><li><p><strong>Agent orchestration.</strong> When to auto-spawn specialized subagents. Security-sensitive code triggers a security-reviewer. Build failures trigger a build-error-resolver. These aren&#8217;t optional. They&#8217;re wired into the rules.</p></li><li><p><strong>Testing standards.</strong> 80% coverage minimum. TDD when requested. xUnit with AAA pattern. <code>WebApplicationFactory&lt;Program&gt;</code> for integration tests.</p></li><li><p><strong>Performance guardrails.</strong> Context window management, research time limits, build troubleshooting protocols.</p></li><li><p><strong>Cross-project memory.</strong> When to query Nexus before asking me to re-explain something. When to check for existing decisions instead of making new ones.</p></li><li><p><strong>Common patterns.</strong> Privacy tags, skeleton project discovery, reusable architectural templates.</p></li></ul><p>The cost: approximately 3,000 tokens of system context. The payoff: the model writes code that matches my standards on the first attempt. No correction cycles. No &#8220;actually, we use FluentValidation, not DataAnnotations.&#8221; No &#8220;please don&#8217;t use console.log.&#8221;</p><h3><strong>Layer 2: Lifecycle hooks (16 hooks across 6 phases)</strong></h3><p>This is where a harness becomes autonomous. Claude Code supports hooks at six lifecycle phases: SessionStart, PreToolUse, PostToolUse, PreCompact, PostCompact, and Stop. I use all six.</p><p><strong>SessionStart (3 hooks):</strong></p><ul><li><p>Memory persistence restores my previous session&#8217;s context: what I was working on, what&#8217;s next, what decisions were made.</p></li><li><p><a href="https://github.com/MCKRUZ/Nexus">Nexus</a> syncs at session start. Nexus is a tool I built, a local-first knowledge graph that spans all my projects (more on this in Layer 6). It loads architectural decisions and dependency maps from all 30+ projects so the model doesn&#8217;t start cold.</p></li><li><p>Nexus session initialization sets up cross-project tracking for the current session.</p></li></ul><p><strong>PreToolUse (3 hooks):</strong></p><ul><li><p>A skill switchboard routes file edits to the appropriate skill based on file type and project context.</p></li><li><p>A strategic compact advisor monitors context usage and suggests compaction before I hit limits, preventing emergency compactions that lose more state.</p></li><li><p>A push safety gate lists commits before any <code>git push</code>, requiring review.</p></li></ul><p><strong>PostToolUse (3 hooks):</strong></p><ul><li><p>Auto-formatting runs after every edit, eliminating style inconsistency.</p></li><li><p>Observation logging records file paths and patterns to session memory.</p></li><li><p>Nexus records architectural patterns and decisions for cross-project learning.</p></li></ul><p><strong>PreCompact &amp; PostCompact (2 hooks):</strong></p><ul><li><p>Before compaction: update MEMORY.md, save reusable patterns to lesson storage.</p></li><li><p>After compaction: re-orient from MEMORY.md and task list so the conversation continues without interruption.</p></li></ul><p><strong>Stop (5 hooks):</strong></p><ul><li><p>Clean MCP subprocess termination.</p></li><li><p>Autonomy throttle (tracks how often Claude stops to check in; if it stops 3+ times in 5 minutes, it pauses and asks what I need instead of continuing on autopilot).</p></li><li><p>Session end persistence.</p></li><li><p>Nexus post-session analysis.</p></li><li><p>Telemetry export for cost and token tracking.</p></li></ul><p>The critical insight: <strong>these hooks run without my intervention.</strong> I don&#8217;t think about memory persistence or auto-formatting or session telemetry. The harness handles it. Every cognitive cycle I&#8217;m not spending on housekeeping is a cycle I&#8217;m spending on the actual problem.</p><h3><strong>Layer 3: Skills (28 domain expertise modules)</strong></h3><p>This is the layer nobody talks about, and it might be the most underrated.</p><p>Claude Code skills are reusable prompt templates, slash commands that load pre-encoded domain expertise into the conversation on demand. Instead of writing a 500-token ad-hoc prompt explaining what you want, you invoke a skill that loads optimized, battle-tested instructions.</p><p>I maintain 28 skills across two tiers: 18 global skills available in every project, and 10 project-specific skills tailored to individual workflows.</p><p><strong>Engineering skills:</strong></p><ul><li><p><code>/tdd-workflow</code> enforces test-driven development: scaffold interfaces, write tests first, implement to pass, verify coverage. No ambiguity about the TDD process.</p></li><li><p><code>/security-review</code> triggers a comprehensive security audit for auth, payment, or identity code. Covers OWASP Top 10, secrets management, and attack surface analysis.</p></li><li><p><code>/shannon</code> is an autonomous AI pentester. White-box security assessments with real exploit execution. Not theoretical. It finds actual vulnerabilities.</p></li><li><p><code>/llm-cost-optimizer</code> audits model selections across call sites and recommends cheaper alternatives by complexity tier. Pays for itself immediately.</p></li><li><p><code>/code-review</code>, <code>/build-fix</code>, <code>/refactor-clean</code> are post-implementation quality gates that catch issues before they compound.</p></li></ul><p><strong>Design and visualization skills:</strong></p><ul><li><p><code>/design</code> covers full brand identity, design tokens, UI styling, and logo generation with 55+ styles.</p></li><li><p><code>/slides</code> creates strategic HTML presentations with Chart.js, design tokens, and contextual slide strategies.</p></li><li><p><code>/visual-explainer</code> generates self-contained HTML pages that visually explain systems, code changes, and data.</p></li><li><p><code>/dashboard-creator</code> builds KPI metric cards, charts, and data visualizations as standalone HTML.</p></li></ul><p><strong>Planning and analysis skills:</strong></p><ul><li><p><code>/deep-plan</code> creates sectionized implementation plans with multi-LLM review and stakeholder interviews.</p></li><li><p><code>/skeptic</code> is one I built myself (<a href="https://github.com/MCKRUZ/skeptic-skill">open source on GitHub</a>). It runs critical analysis that pokes holes in plans, surfaces hidden complexity, and challenges assumptions before I commit to an approach. This one exists because AI agents have an agreeableness problem. You describe an idea and the model tells you it&#8217;s brilliant, then helps you build it. Three days later you realize the approach had an obvious flaw that a skeptical colleague would have caught in five minutes. <code>/skeptic</code> is that colleague. It forces the model to argue <em>against</em> the plan before I greenlight it, and it&#8217;s saved me from more bad architectural decisions than any other single tool in this harness.</p></li><li><p><code>/functional-design</code> handles end-to-end UI/UX creation from functional spec to working code.</p></li></ul><p>The context window angle on skills is easy to miss. Without a skill, I type out detailed instructions every time, and the model still pulls its punches or misses nuance. That means follow-up corrections. With a skill, I type three words and the model loads 2,000 tokens of pre-optimized instructions. Comprehensive and correct from the first pass. Zero follow-up cycles. Skills convert ad-hoc prompting (variable quality, frequent corrections) into encoded expertise (consistent quality, zero corrections).</p><h3><strong>Layer 4: Plugins (5 active)</strong></h3><p>Claude Code&#8217;s plugin ecosystem is young, but the right plugins dramatically extend capability:</p><ul><li><p><strong>context-mode</strong> is the biggest context window optimizer in my stack. Instead of dumping raw command output (build logs, git history, file contents) directly into the context window, context-mode sandboxes execution and indexes the output in a local SQLite FTS5 database. Only summaries enter my conversation. A 500-line <code>git log</code> that would normally consume ~8,000 tokens? Indexed and searchable for a few hundred. A build failure with 200 lines of stack trace? Summarized to the 3 relevant error lines. Over a session, this compounds fast.</p></li><li><p><strong>deep-plan</strong> provides sectionized implementation planning with multi-LLM review. Turns vague feature requests into structured, reviewable plans before a line of code is written.</p></li><li><p><strong>deep-implement</strong> runs TDD-oriented implementation from deep-plan sections. Writes tests first, implements to pass, then reviews.</p></li><li><p><strong>deep-project</strong> decomposes high-level project requirements into scoped planning units.</p></li><li><p><strong>code-simplifier</strong> handles post-implementation cleanup. Reviews changed code for reuse opportunities and unnecessary complexity.</p></li></ul><h3><strong>Layer 5: MCP server aggregation (Bifrost)</strong></h3><p>MCP (Model Context Protocol) servers give Claude Code access to external tools, from APIs to documentation indexes to video analytics, through a standardized interface. The problem: each MCP server exposes its own set of tool definitions, and every tool definition consumes context window tokens. Run five MCP servers with 3-5 tools each, and you&#8217;re spending thousands of tokens just on tool schemas before you&#8217;ve asked a single question.</p><p><strong>Bifrost</strong> solves this by acting as an HTTP gateway that aggregates multiple MCP servers behind a single endpoint. It runs on a dedicated machine in my network and proxies requests to backend servers:</p><ul><li><p><strong>context7</strong> provides library documentation and code examples on demand. When Claude needs to reference a framework&#8217;s API, it queries context7 instead of burning tokens on web searches or hallucinating method signatures. This is the one I lean on hardest. Accurate docs in the context window, not guesswork.</p></li><li><p><strong>sequential-thinking</strong> enables structured multi-step reasoning for complex architectural decisions. Forces the model to decompose a problem before committing to an approach, rather than jumping straight to code.</p></li><li><p><strong>github</strong> provides direct GitHub API access for PR management, issue tracking, and repository operations without leaving the conversation.</p></li><li><p><strong>firecrawl</strong> handles web scraping and content extraction. When I need to pull in a competitor&#8217;s documentation, an API reference, or a technical article, firecrawl fetches and cleans it without me alt-tabbing to a browser.</p></li><li><p><strong>youtube</strong> provides video analytics and transcript extraction for competitive research and content analysis.</p></li></ul><p>Two additional MCP servers run standalone: <strong>Nexus</strong> (cross-project memory graph) and <strong>CMEM</strong> (session memory). These stay direct because they&#8217;re lightweight, local-only operations that don&#8217;t benefit from gateway aggregation.</p><p>That&#8217;s 8 MCP server endpoints total: 5 through Bifrost, 2 standalone, plus Bifrost itself. From the model&#8217;s perspective, it sees 3 MCP connections. From my perspective, it&#8217;s 3 processes to manage instead of 8.</p><p>The token savings come from two places. First, Bifrost consolidates all five servers behind one HTTP endpoint with one schema negotiation instead of five, reducing tool definition overhead in the context window. Second, and more importantly, these MCP servers return focused results instead of raw output. A context7 documentation query returns the exact function signature, parameters, and a usage example in 200-400 tokens. The alternative? A web search that dumps 3,000-6,000 tokens of HTML, ads, and irrelevant content into the context window. Firecrawl returns clean markdown instead of raw DOM. These MCP servers return what the model needs, not everything the source contains.</p><p>The bigger point: <strong>MCP servers are how you give an agent real-world reach without bloating its context.</strong> The harness controls what enters the context window, and Bifrost is the gatekeeper.</p><h3><strong>Layer 6: Cross-project memory (Nexus + CMEM)</strong></h3><p>This is the layer most people don&#8217;t realize they&#8217;re missing.</p><p><strong><a href="https://github.com/MCKRUZ/Nexus">Nexus</a></strong> is a tool I built: a local-first, encrypted knowledge graph that spans all my projects. When I make an architectural decision in Project A, Nexus records it. When I&#8217;m working on Project B and face a similar decision, Nexus surfaces the prior art. It tracks:</p><ul><li><p>Architectural decisions and their rationale</p></li><li><p>Code patterns and where they&#8217;re used</p></li><li><p>Cross-project dependencies and potential conflicts</p></li><li><p>Infrastructure notes (SSH configs, deployment targets, service accounts)</p></li></ul><p><strong>CMEM</strong> (session memory) provides semantic search across past conversations. When I worked on a similar problem three weeks ago, CMEM surfaces the relevant context without me re-explaining it.</p><p>Without cross-project memory, every session starts cold. The model reads files it read yesterday, asks questions I&#8217;ve answered before, and explores architecture it&#8217;s already mapped. With Nexus and CMEM, the model loads synced context at session start and picks up where we left off. The first task of every session starts productive instead of exploratory.</p><h3><strong>Layer 7: Observability (Langfuse + status line)</strong></h3><p>You can&#8217;t optimize what you don&#8217;t measure. My harness exports every session to Langfuse for cost and token tracking. A custom status line shows project name, model, context usage percentage, and remaining capacity in real time.</p><p>This might seem like a nice-to-have. It&#8217;s not. When you can see that a research task consumed 45% of your context window on raw command output, you know exactly where to optimize. When you can compare token-per-task costs across sessions, you can make data-driven decisions about which plugins and rules are pulling their weight.</p><p>The status line is also a context window guardian. Seeing &#8220;Context: 72% | ~56K remaining&#8221; in real time changes how you work. You don&#8217;t issue a massive <code>git log --all</code> when you can see you&#8217;re at 70% capacity. You reach for context-mode&#8217;s sandboxed execution instead. Observability turns unconscious context waste into conscious context management.</p><h2><strong>The token math</strong></h2><p>Here&#8217;s what this looks like in practice. These are representative estimates from my Langfuse telemetry across typical coding sessions. Individual sessions vary, but the pattern is consistent.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!D10P!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafde6a4c-42b1-41fd-9618-12afbf78e19b_1140x718.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!D10P!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafde6a4c-42b1-41fd-9618-12afbf78e19b_1140x718.png 424w, https://substackcdn.com/image/fetch/$s_!D10P!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafde6a4c-42b1-41fd-9618-12afbf78e19b_1140x718.png 848w, https://substackcdn.com/image/fetch/$s_!D10P!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafde6a4c-42b1-41fd-9618-12afbf78e19b_1140x718.png 1272w, https://substackcdn.com/image/fetch/$s_!D10P!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafde6a4c-42b1-41fd-9618-12afbf78e19b_1140x718.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!D10P!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafde6a4c-42b1-41fd-9618-12afbf78e19b_1140x718.png" width="1140" height="718" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/afde6a4c-42b1-41fd-9618-12afbf78e19b_1140x718.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:718,&quot;width&quot;:1140,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:83426,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://matthewkruczek.substack.com/i/192637819?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafde6a4c-42b1-41fd-9618-12afbf78e19b_1140x718.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!D10P!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafde6a4c-42b1-41fd-9618-12afbf78e19b_1140x718.png 424w, https://substackcdn.com/image/fetch/$s_!D10P!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafde6a4c-42b1-41fd-9618-12afbf78e19b_1140x718.png 848w, https://substackcdn.com/image/fetch/$s_!D10P!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafde6a4c-42b1-41fd-9618-12afbf78e19b_1140x718.png 1272w, https://substackcdn.com/image/fetch/$s_!D10P!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fafde6a4c-42b1-41fd-9618-12afbf78e19b_1140x718.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>The combined effect: my sessions run at roughly 70-80% productive work ratio compared to the ~45% I see in default configurations. Compactions drop from 4-6 per session to 1-2. The model gets things right on the first attempt instead of the third. And I spend my time thinking about the actual problem instead of re-explaining my project.</p><h2><strong>Same prompt, different worlds</strong></h2><p>Let me make this concrete. Here&#8217;s a real prompt I might give Claude Code:</p><blockquote><p>&#8220;Refactor the authentication middleware to use the Result pattern and add FluentValidation&#8221;</p></blockquote><h3><strong>Without a harness (default Claude Code):</strong></h3><ol><li><p><strong>Claude asks clarifying questions</strong> (500 tokens): &#8220;What framework? What&#8217;s the project structure? What&#8217;s the Result pattern you&#8217;re using?&#8221;</p></li><li><p><strong>Exploration phase</strong> (20,000 tokens): Reads 10-15 files to understand the codebase, architecture, existing patterns</p></li><li><p><strong>First implementation attempt</strong> (1,500 tokens): Writes code with <code>try/catch</code> and <code>DataAnnotations</code>. Reasonable, but wrong for this codebase.</p></li><li><p><strong>User correction</strong> (800 tokens): &#8220;We use FluentValidation, not DataAnnotations. And we use <code>Result&lt;T&gt;</code> with error discriminated unions.&#8221;</p></li><li><p><strong>Second attempt</strong> (1,500 tokens): Closer, but uses mutable patterns and <code>console.log</code> for debugging</p></li><li><p><strong>User correction</strong> (600 tokens): &#8220;Immutable only. No console.log. Use structured logging.&#8221;</p></li><li><p><strong>Third attempt</strong> (1,500 tokens): Finally correct</p></li><li><p><strong>No security review triggered</strong></p></li><li><p><strong>No tests written</strong></p></li></ol><p><strong>Total: ~26,000 tokens. Three iterations. No tests. No security review.</strong></p><h3><strong>With my harness:</strong></h3><ol><li><p><strong>Rules pre-loaded</strong>: Claude already knows FluentValidation, <code>Result&lt;T&gt;</code>, immutability requirements, structured logging, security standards</p></li><li><p><strong>Nexus syncs</strong> (2,000 tokens): Project architecture, existing middleware patterns, and dependency graph loaded</p></li><li><p><strong>First implementation</strong> (1,500 tokens): Correct patterns from the start. Immutable, FluentValidation, Result, structured logging.</p></li><li><p><strong>Security reviewer auto-spawns</strong> (rules detect auth code): Reviews for JWT validation, CSRF, token storage</p></li><li><p><strong>Auto-format runs</strong>: Code style consistent without manual review</p></li><li><p><strong>Build output sandboxed</strong> by context-mode: a few hundred tokens instead of thousands</p></li></ol><p><strong>Total: ~10,000 tokens. One iteration. Security reviewed. Standards enforced.</strong></p><p>That&#8217;s not a 10% improvement. That&#8217;s a fundamentally different relationship with your context window.</p><h2><strong>Context window optimization: the throughline</strong></h2><p>If you&#8217;ve been counting, every layer of this harness optimizes the same scarce resource. That&#8217;s not a coincidence. It&#8217;s the design principle.</p><p><strong>Rules</strong> front-load 3,000 tokens to prevent thousands more in correction cycles. <strong>Hooks</strong> automate state management that would otherwise consume context with manual &#8220;remember what we were doing&#8221; prompts. <strong>Skills</strong> convert variable-quality ad-hoc instructions into consistent, pre-optimized prompts. <strong>context-mode</strong> keeps raw tool output out of the window entirely. <strong>Bifrost</strong> routes queries to MCP servers that return focused results instead of raw HTML. <strong>Memory</strong> eliminates the cold-start exploration that consumes 10-20% of every default session. <strong>Observability</strong> makes all of this measurable so you know what&#8217;s working.</p><p>The context window is the fundamental constraint of agentic coding. Every token spent on overhead is a token not spent on reasoning about your actual problem. Every unnecessary compaction is lost state and degraded continuity. Every correction cycle is the model doing the same work twice.</p><p>A harness doesn&#8217;t make the model faster or smarter. It makes the context window deeper. And depth, the ability to sustain complex, multi-step reasoning without losing state, is what separates a coding assistant from a coding agent.</p><h2><strong>This is the new skill gap</strong></h2><p>Here&#8217;s why this matters beyond my personal setup.</p><p>In every previous technology wave, the differentiator was knowledge of the technology itself. Know Java better than the next person, ship faster. Know Kubernetes better, deploy more reliably.</p><p>With AI coding agents, the technology is the same for everyone. We all have access to the same Claude, the same GPT, the same Gemini. The model isn&#8217;t the differentiator.</p><p><strong>The harness is.</strong></p><p>The engineer who spends a week configuring governance rules, building skills for their recurring workflows, and optimizing context consumption will outperform the engineer with default settings for the next two years. Every session. Every task. Compounding daily.</p><p>The gap compounds in both speed and quality. The harnessed engineer gets security reviews triggered automatically. They get consistent code patterns without thinking about it. Their sessions don&#8217;t break every 30 minutes from compaction. Their model has access to documentation through MCP tools instead of hallucinating API signatures from training data.</p><p>This is also why I keep arguing that <a href="https://file+.vscode-resource.vscode-cdn.net/c%3A/Users/kruz7/OneDrive/Documents/Code%20Repos/MCKRUZ/matthewkruczek-ai/blog/drafts/agent-harnesses-fewer-layers.html">the agent harness is the architecture</a>. It&#8217;s not infrastructure you bolt on after the fact. It&#8217;s the primary lever for engineering productivity in an AI-native workflow.</p><h2><strong>What to do this week</strong></h2><p>You don&#8217;t need 16 hooks, 28 skills, and 5 plugins on day one. Start with the highest-leverage layers, the ones that save the most context for the least effort:</p><p><strong>Start with rules.</strong> Create <code>~/.claude/rules/coding-style.md</code> and encode your team&#8217;s three most-violated coding standards. This single file will eliminate the most common correction cycles immediately.</p><p><strong>Add a CLAUDE.md to your project.</strong> Document your stack, your patterns, and your conventions. This is the project onboarding document you wish every new hire got on day one, except now your AI agent reads it every session. It&#8217;s the highest-leverage single file in your entire repository.</p><p><strong>Install context-mode.</strong> If you use Claude Code for any research or debugging, this plugin will be the single biggest context window optimizer. Raw command output flooding your context window is the #1 source of unnecessary compactions.</p><p><strong>Build your first skill.</strong> Take the prompt you type most often, the one with specific formatting requirements, voice guidelines, or domain conventions, and turn it into a skill. It takes 30 minutes. It saves correction cycles forever.</p><p><strong>Set up memory persistence.</strong> Even a simple PreCompact hook that prompts the model to update a MEMORY.md file will dramatically improve cross-session continuity. You shouldn&#8217;t have to re-explain your project every conversation.</p><p><strong>Measure everything.</strong> You can&#8217;t optimize a context window you don&#8217;t observe. Start tracking token consumption per task. The patterns will be obvious, and they&#8217;ll tell you exactly which layer to build next.</p><p>The model is the engine. The context window is the fuel tank. The harness determines how far you go on every drop.</p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://matthewkruczek.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://matthewkruczek.substack.com/subscribe?"><span>Subscribe now</span></a></p><p></p><div><hr></div><p><em>Matthew Kruczek is Managing Director at EY, leading Microsoft domain initiatives within Digital Engineering. Connect with Matthew on <a href="https://www.linkedin.com/in/matthew-kruczek/">LinkedIn</a> to discuss agent harness architecture and context optimization for your engineering organization.</em></p><h2><strong>References</strong></h2><ol><li><p>Kruczek, M. &#8220;Agent Harnesses Don&#8217;t Need More Layers. They Need Fewer.&#8221; matthewkruczek.ai, March 17, 2026. <a href="https://file+.vscode-resource.vscode-cdn.net/c%3A/Users/kruz7/OneDrive/Documents/Code%20Repos/MCKRUZ/matthewkruczek-ai/blog/drafts/agent-harnesses-fewer-layers.html">matthewkruczek.ai</a></p></li><li><p>Anthropic. &#8220;Claude Code: Hooks.&#8221; Anthropic Documentation, 2026. <a href="https://docs.anthropic.com/en/docs/claude-code/hooks">docs.anthropic.com</a></p></li><li><p>Anthropic. &#8220;Claude Code: CLAUDE.md.&#8221; Anthropic Documentation, 2026. <a href="https://docs.anthropic.com/en/docs/claude-code/claude-md">docs.anthropic.com</a></p></li><li><p>mksglu. &#8220;Context-Mode Plugin for Claude Code.&#8221; GitHub, 2026. <a href="https://github.com/mksglu/context-mode">github.com</a></p></li><li><p>Kruczek, M. &#8220;Progressive Disclosure for MCP Servers.&#8221; matthewkruczek.ai, 2026. <a href="https://file+.vscode-resource.vscode-cdn.net/c%3A/Users/kruz7/OneDrive/Documents/Code%20Repos/MCKRUZ/matthewkruczek-ai/blog/drafts/progressive-disclosure-mcp-servers.html">matthewkruczek.ai</a></p></li><li><p>Kruczek, M. &#8220;Context Engineering for Enterprise AI.&#8221; matthewkruczek.ai, 2026. <a href="https://file+.vscode-resource.vscode-cdn.net/c%3A/Users/kruz7/OneDrive/Documents/Code%20Repos/MCKRUZ/matthewkruczek-ai/blog/drafts/context-engineering-enterprise-ai.html">matthewkruczek.ai</a></p></li><li><p>Anthropic. &#8220;Claude Code: Model Context Protocol.&#8221; Anthropic Documentation, 2026. <a href="https://docs.anthropic.com/en/docs/claude-code/mcp">docs.anthropic.com</a></p></li><li><p>Bifrost. &#8220;MCP Gateway for Claude Code.&#8221; GitHub, 2026. <a href="https://github.com/BifrostMCP/bifrost">github.com</a></p></li><li><p>Anthropic. &#8220;Claude Code: Skills.&#8221; Anthropic Documentation, 2026. <a href="https://docs.anthropic.com/en/docs/claude-code/skills">docs.anthropic.com</a></p></li></ol>]]></content:encoded></item><item><title><![CDATA[The Missing Layer: Why Your AI Agents Need a Package Manager]]></title><description><![CDATA[If you only have a minute, here&#8217;s what you need to know.]]></description><link>https://matthewkruczek.substack.com/p/the-missing-layer-why-your-ai-agents</link><guid isPermaLink="false">https://matthewkruczek.substack.com/p/the-missing-layer-why-your-ai-agents</guid><dc:creator><![CDATA[Matthew Kruczek]]></dc:creator><pubDate>Mon, 23 Mar 2026 16:01:03 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!sQdv!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79cf3634-b98f-4c1e-8f03-057b6b30af5c_1200x624.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>If you only have a minute, here&#8217;s what you need to know.</em></p><ul><li><p>A 5-person team using AI coding agents today has 5 divergent agent configurations. A 50-person team has 50. Nobody is managing this.</p></li><li><p>Microsoft&#8217;s open-source APM (Agent Package Manager) treats agent configuration with the same rigor we&#8217;ve applied to code dependencies for decades: declare once in <code>apm.yml</code>, resolve transitively, lock versions, reproduce everywhere.</p></li><li><p>APM manages seven primitives that define how agents work in your codebase: instructions, skills, prompts, agents, hooks, plugins, and MCP servers. It compiles a single manifest into native formats for Copilot, Claude Code, Cursor, and OpenCode.</p></li><li><p>Supply chain security for agent configurations is already a concern. APM includes <code>apm audit</code> for detecting hidden Unicode vulnerabilities and blocks compromised packages before agents access them.</p></li><li><p>Agent configuration is infrastructure, not a personal preference. Organizations that treat it as such will compound their advantages the same way teams with disciplined dependency management outperformed those without.</p></li></ul><div><hr></div><h2><strong>The dependency nobody is managing</strong></h2><p>Here&#8217;s something that should bother every engineering leader deploying AI coding agents: your developers are all using different ones.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://matthewkruczek.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Context Engineering! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>Not different models. Different <em>configurations</em>. Different instructions. Different skills. Different MCP servers. Different hooks. Different rules about what the agent should and shouldn&#8217;t do.</p><p>One developer has a carefully tuned CLAUDE.md with coding standards, security guardrails, and domain-specific knowledge. The developer sitting next to them has a bare default. A third copied a configuration from a blog post three months ago and hasn&#8217;t updated it since. Your newest hire has nothing at all.</p><p>This is the dependency management problem of 2015, replaying in fast-forward. We&#8217;ve seen this movie before. And Microsoft just shipped an open-source tool that solves it the same way npm and NuGet solved it for code: APM (Agent Package Manager). One manifest. Transitive resolution. Lock files. Reproducible agent configurations across your entire org.</p><p>Before I walk through what APM does and why it matters, it&#8217;s worth understanding the pattern it&#8217;s built on. This isn&#8217;t a new problem. It&#8217;s a solved problem being applied to a new domain.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!sQdv!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79cf3634-b98f-4c1e-8f03-057b6b30af5c_1200x624.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!sQdv!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79cf3634-b98f-4c1e-8f03-057b6b30af5c_1200x624.png 424w, https://substackcdn.com/image/fetch/$s_!sQdv!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79cf3634-b98f-4c1e-8f03-057b6b30af5c_1200x624.png 848w, https://substackcdn.com/image/fetch/$s_!sQdv!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79cf3634-b98f-4c1e-8f03-057b6b30af5c_1200x624.png 1272w, https://substackcdn.com/image/fetch/$s_!sQdv!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79cf3634-b98f-4c1e-8f03-057b6b30af5c_1200x624.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!sQdv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79cf3634-b98f-4c1e-8f03-057b6b30af5c_1200x624.png" width="1200" height="624" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/79cf3634-b98f-4c1e-8f03-057b6b30af5c_1200x624.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:624,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1053868,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://matthewkruczek.substack.com/i/191880334?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79cf3634-b98f-4c1e-8f03-057b6b30af5c_1200x624.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!sQdv!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79cf3634-b98f-4c1e-8f03-057b6b30af5c_1200x624.png 424w, https://substackcdn.com/image/fetch/$s_!sQdv!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79cf3634-b98f-4c1e-8f03-057b6b30af5c_1200x624.png 848w, https://substackcdn.com/image/fetch/$s_!sQdv!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79cf3634-b98f-4c1e-8f03-057b6b30af5c_1200x624.png 1272w, https://substackcdn.com/image/fetch/$s_!sQdv!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F79cf3634-b98f-4c1e-8f03-057b6b30af5c_1200x624.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><h2><strong>We solved this problem 15 years ago (for code)</strong></h2><p>Software engineering went through an identical maturation cycle with code dependencies. The timeline is instructive.</p><p>In the early days, developers managed libraries by hand. You&#8217;d download a JAR file, drop it in a /lib folder, and pray that everyone on the team had the same version. It worked for small teams. It collapsed at scale. Version conflicts. Missing transitive dependencies. The infamous &#8220;works on my machine&#8221; problem that consumed entire sprints.</p><p>Then package managers arrived. npm for JavaScript. pip for Python. Maven for Java. NuGet for .NET. Cargo for Rust. Each brought the same core innovation: declare your dependencies in a manifest, resolve them automatically, lock versions for reproducibility. package.json. requirements.txt. Cargo.toml. Different ecosystems, identical principle.</p><p>The effect wasn&#8217;t incremental. It was foundational. Package managers didn&#8217;t just save time. They made entire categories of problems disappear. They enabled open-source ecosystems to scale from dozens of libraries to millions. They made CI/CD pipelines possible by guaranteeing build reproducibility. They turned &#8220;dependency management&#8221; from a skill into a solved problem.</p><p>AI agent configuration is sitting right where code dependencies sat before package managers arrived. Manual setup. Configuration drift. No version pinning. No transitive resolution. No reproducibility guarantees. No security scanning.</p><p>The only difference is the speed. Code dependency management took a decade to mature. Agent configuration needs to get there in months, because the adoption curve is steeper and the blast radius of a misconfigured agent is larger than a mismatched library version.</p><h2><strong>Enter APM</strong></h2><p>Microsoft&#8217;s Agent Package Manager is an open-source project (MIT licensed, 710+ stars, 33 releases as of March 2026) that applies the package manager pattern to AI agent configuration. The creator, Daniel Meppiel, a Software Global Black Belt at Microsoft, built it on a straightforward premise: if package.json works for code dependencies, apm.yml should work for agent dependencies.</p><p>The comparison is more than metaphorical. APM implements the same lifecycle that made code package managers transformative:</p><p><strong>Declare.</strong> Your project&#8217;s apm.yml specifies what your agents need. Not how to configure them manually, but what packages of configuration to install.</p><pre><code><code>name: my-enterprise-project
version: 1.0.0
dependencies:
  apm:
    - org/coding-standards
    - org/security-guardrails
    - org/api-review-skill#v2.1
    - anthropics/skills/skills/frontend-design
</code></code></pre><p><strong>Resolve.</strong> <code>apm install</code> pulls packages from any git host (GitHub, GitLab, Bitbucket, Azure DevOps, GitHub Enterprise), resolves transitive dependencies, and places files in the correct directories.</p><p><strong>Lock.</strong> apm.lock.yaml pins every dependency to an exact commit. The same lock file produces the same agent configuration on every machine, every time.</p><p><strong>Compile.</strong> A single manifest generates native configuration files for multiple tools. AGENTS.md for Copilot. CLAUDE.md for Claude Code. .cursor/rules/ for Cursor. One source of truth, many deployment targets.</p><p><strong>Distribute.</strong> <code>apm pack</code> bundles resolved dependencies into portable artifacts that work without APM, Python, or network access. This covers CI pipelines, air-gapped environments, and dev container setups.</p><p>This isn&#8217;t a prototype. APM is at v0.8.4 with 18 contributors, 772 commits, and a GitHub Action for CI integration. It&#8217;s built on established open standards: the AGENTS.md specification (a Linux Foundation AAIF founding project alongside MCP), the Agent Skills framework, and the Model Context Protocol.</p><h2><strong>Seven primitives, one manifest</strong></h2><p>APM manages seven types of agent configuration that together define how AI coding agents operate in your codebase:</p><p><strong>Instructions</strong> are coding standards and guardrails scoped to file patterns. A Python standards instruction applies only to \<em>\</em>/\*.py files. A security policy applies everywhere. These replace the ad-hoc &#8220;paste your coding standards into the system prompt&#8221; approach that most teams use today.</p><p><strong>Skills</strong> are reusable AI capabilities: code reviewers, form builders, deployment assistants, security scanners. If you&#8217;ve read my previous article on Agent Skills as the missing link, this is the distribution mechanism that was absent. APM makes skills installable, composable, and version-pinned instead of manually copied between repositories.</p><p><strong>Prompts</strong> are slash commands for common tasks. /review, /deploy, /test. Standardized across the team instead of each developer inventing their own.</p><p><strong>Agents</strong> are specialized personas with defined roles, tools, and behavioral boundaries. Your security review agent. Your architecture advisor. Your documentation generator.</p><p><strong>Hooks</strong> are lifecycle event handlers that fire on specific triggers: before a commit, after a file edit, on session start. These automate quality gates and organizational workflows.</p><p><strong>Plugins</strong> are pre-packaged bundles that combine multiple primitives into a single installable unit. Think of them as meta-packages.</p><p><strong>MCP Servers</strong> are external tool integrations that give agents access to databases, APIs, file systems, and other services through the Model Context Protocol.</p><p>Without APM, configuring all seven of these for a new developer on your team is a multi-hour manual process that nobody documents and everyone does differently. With APM, it&#8217;s <code>git clone &amp;&amp; apm install</code>.</p><h2><strong>The problem that&#8217;s hiding in plain sight</strong></h2><p>Most organizations don&#8217;t realize they have an agent configuration problem because they haven&#8217;t tried to standardize yet. The moment you do, the scale of the issue becomes obvious.</p><p>Imagine a financial services firm with 200 developers using AI coding agents. Some teams use Claude Code. Others use Copilot. A few use Cursor. Each team has its own coding standards, security requirements, and compliance guardrails. Some teams have encoded these into their agent configurations. Most haven&#8217;t.</p><p>Without a dependency manager, you&#8217;re left with two options. Option one: a wiki page titled &#8220;How to Set Up Your AI Agent&#8221; that&#8217;s perpetually out of date and that new hires follow maybe half of. Option two: a senior engineer spends an afternoon with each new team member walking them through the setup, which means the configuration is as good as that engineer&#8217;s memory on that particular day.</p><p>Neither option scales. Neither option is auditable. Neither option gives you confidence that the agent writing code for your regulated application has the same security guardrails as the agent that passed your last compliance review.</p><p>APM&#8217;s lock file changes this equation. apm.lock.yaml is a diffable, version-controlled artifact that proves every developer on your team is running the same agent configuration. When your compliance team asks &#8220;how do we know the AI agents are following our security standards?&#8221;, you show them the lock file.</p><h2><strong>Supply chain security for agent configuration</strong></h2><p>Here&#8217;s where it gets serious. AI agent configurations aren&#8217;t just preferences. They&#8217;re instructions that shape what code gets written, what security patterns get followed, and what guardrails get applied. A compromised agent configuration is a supply chain attack on your entire codebase.</p><p>APM includes <code>apm audit</code>, which scans packages for hidden Unicode vulnerabilities and blocks compromised configurations before they reach your agents. This is the agent equivalent of npm audit or GitHub&#8217;s Dependabot alerts. It&#8217;s table stakes for any organization that takes software supply chain security seriously.</p><p>CIO.com recently identified AI agent versioning as &#8220;the CIO&#8217;s next big challenge,&#8221; noting that &#8220;a minor API update could significantly alter an agent&#8217;s behavior, even if the agent&#8217;s core logic remains unchanged.&#8221; The article recommended treating agent versioning as &#8220;a first-class discipline&#8221; alongside traditional software release management.</p><p>APM makes that recommendation actionable. Version pinning. Lock files. Auditing. Reproducible builds. These aren&#8217;t new concepts. They&#8217;re proven infrastructure patterns applied to a new domain.</p><h2><strong>Where this fits in the agent-first enterprise</strong></h2><p>Throughout this series, I&#8217;ve argued that the organizations seeing transformational results with AI aren&#8217;t just using agent tools. They&#8217;re building infrastructure around agent capabilities.</p><p>Skills gave us composable, portable organizational knowledge. MCP gave us a standard protocol for connecting agents to external tools. Progressive disclosure solved the context window scaling problem.</p><p>APM is the layer that was missing between all of these. It&#8217;s the composition and governance infrastructure that answers questions the other standards don&#8217;t address:</p><ul><li><p>Which skills should this project use? (Declared in apm.yml)</p></li><li><p>What version of those skills? (Pinned in apm.lock.yaml)</p></li><li><p>Are all developers using the same setup? (Reproducible via <code>apm install</code>)</p></li><li><p>Are the configurations safe? (Verified via <code>apm audit</code>)</p></li><li><p>How do we distribute this to new team members? (Automated via <code>apm pack</code>)</p></li></ul><p>The progression looks like this: first you adopt agents, then you customize them with skills and MCP servers, then you realize you need a way to manage and distribute those customizations consistently. APM is what you reach for at that third step.</p><h2><strong>What to do this week</strong></h2><p><strong>If you&#8217;re already using AI coding agents in your organization:</strong></p><p>Start by auditing what you have. How many different agent configurations exist across your teams? Are coding standards encoded in those configurations? Security guardrails? If you can&#8217;t answer these questions, you&#8217;ve confirmed the problem.</p><p>Install APM. It takes less than a minute. On Windows: <code>irm https://aka.ms/apm-windows | iex</code>. On Mac/Linux: <code>curl -sSL https://aka.ms/apm-unix | sh</code>. Run <code>apm init</code> in a project to see the manifest structure.</p><p><strong>This month:</strong></p><p>Pick one team as a pilot. Have them declare their existing agent configuration in apm.yml, including their coding standards, security rules, and any skills or MCP servers they&#8217;ve configured manually. Install from a shared git repository so the entire team gets the same setup.</p><p>Track two things: time saved on new developer onboarding, and the number of configuration inconsistencies you discover during the migration. Both numbers will surprise you.</p><p><strong>This quarter:</strong></p><p>Build an organizational package library. Your coding standards. Your security guardrails. Your compliance requirements. Your domain-specific skills. Publish them to internal git repositories that any team can install from. Establish a review process for changes to these packages, the same way you review changes to shared libraries.</p><p>Integrate <code>apm audit</code> into your CI pipeline. Make agent configuration a first-class artifact alongside your code.</p><h2><strong>The infrastructure maturity curve</strong></h2><p>Every technology goes through the same maturation arc. First comes capability: the thing works. Then adoption: people start using it. Then chaos: everyone uses it differently. Then infrastructure: tooling emerges to manage the chaos. Then standardization: the tooling becomes an expected part of the stack.</p><p>AI coding agents are somewhere between chaos and infrastructure. The capability is proven. Adoption is accelerating (96% of enterprise IT leaders plan to expand AI agent use in the next 12 months). The chaos is real but largely unacknowledged. The infrastructure is arriving.</p><p>APM is one of the first serious attempts to move agent configuration from artisanal to industrial. It won&#8217;t be the last. But it&#8217;s built on the right foundations (open standards, proven patterns, multi-tool support), and it&#8217;s solving a problem that gets worse with every developer you add to your team.</p><p>The organizations that figure out agent infrastructure early will compound their advantages over those that don&#8217;t. That&#8217;s not speculation. It&#8217;s exactly what happened with code dependency management, with CI/CD, with containerization, with every infrastructure layer that made software engineering more reproducible and reliable.</p><p>Agent configuration is infrastructure. Treat it accordingly.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://matthewkruczek.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Context Engineering! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[You're Measuring Agentic AI Wrong: The Three-Layer Framework Leaders Actually Need]]></title><description><![CDATA[If you only have a minute, here&#8217;s what you need to know.]]></description><link>https://matthewkruczek.substack.com/p/youre-measuring-agentic-ai-wrong</link><guid isPermaLink="false">https://matthewkruczek.substack.com/p/youre-measuring-agentic-ai-wrong</guid><dc:creator><![CDATA[Matthew Kruczek]]></dc:creator><pubDate>Thu, 19 Mar 2026 13:28:44 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!_MeP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb835ac0-d1f7-4e4a-a5e3-bab8e8b4c948_1200x624.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>If you only have a minute, here&#8217;s what you need to know.</em></p><ul><li><p>Agentic AI doesn&#8217;t just automate tasks. It makes decisions within processes. Measuring it with the same KPIs you used for RPA will give you numbers that move without telling you anything useful.</p></li><li><p>Most enterprises are sitting at Layer 1: operational metrics like FTE saved and cycle time reduction. These matter, but they&#8217;re the floor, not the ceiling.</p></li><li><p>Layer 2 metrics measure decision quality: how often the agent made the right call, how often humans overrode it, and where confidence broke down. This is where most organizations have a measurement gap today.</p></li><li><p>Layer 3 metrics measure adaptive capacity: whether the agent is expanding what the process can do, not just running the existing process faster. This is where transformational value lives.</p></li><li><p>The biggest measurement mistake leaders make is declaring success when Layer 1 numbers improve, before checking whether Layers 2 and 3 are even being tracked.</p></li></ul><p>Nearly every leader I talk to right now has some version of the same conversation. Their team has introduced an AI agent into a core process. Something is clearly happening. The numbers are moving. But when the business asks &#8220;what&#8217;s the impact?&#8221;, the answer sounds thinner than the experience feels.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://matthewkruczek.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Context Engineering! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>That gap between what&#8217;s happening and what the metrics are capturing is not an accident. It&#8217;s structural.</p><p>Agentic AI is not automation. Traditional automation executes a predetermined sequence of steps. An agent evaluates context, makes judgment calls, handles exceptions, and adapts to conditions the original process designer never anticipated. When you put an agent into a business process, you are not installing a faster conveyor belt. You are introducing a new decision-maker.</p><p>Most enterprises are still measuring the conveyor belt.</p><h2><strong>The problem with your current dashboard</strong></h2><p>When organizations first deploy AI agents, the metrics that surface naturally are the ones already on their dashboards: cycle time, task volume, FTE hours, cost per transaction, error rate. These numbers are real and they matter. But they were designed to measure deterministic processes, where the only question is how fast and how reliably the sequence runs.</p><p>Agents introduce non-determinism. They don&#8217;t always take the same path. They encounter novel situations and handle them in ways the process never specified. They can escalate appropriately, fail silently, or make calls that are locally correct but strategically wrong.</p><p>None of that is visible in your existing KPIs.</p><p>The organizations that are ahead on this have built a three-layer measurement framework. Each layer answers a different question.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!_MeP!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb835ac0-d1f7-4e4a-a5e3-bab8e8b4c948_1200x624.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!_MeP!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb835ac0-d1f7-4e4a-a5e3-bab8e8b4c948_1200x624.png 424w, https://substackcdn.com/image/fetch/$s_!_MeP!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb835ac0-d1f7-4e4a-a5e3-bab8e8b4c948_1200x624.png 848w, https://substackcdn.com/image/fetch/$s_!_MeP!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb835ac0-d1f7-4e4a-a5e3-bab8e8b4c948_1200x624.png 1272w, https://substackcdn.com/image/fetch/$s_!_MeP!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb835ac0-d1f7-4e4a-a5e3-bab8e8b4c948_1200x624.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!_MeP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb835ac0-d1f7-4e4a-a5e3-bab8e8b4c948_1200x624.png" width="1200" height="624" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/db835ac0-d1f7-4e4a-a5e3-bab8e8b4c948_1200x624.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:624,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:408985,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://matthewkruczek.substack.com/i/191470290?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb835ac0-d1f7-4e4a-a5e3-bab8e8b4c948_1200x624.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!_MeP!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb835ac0-d1f7-4e4a-a5e3-bab8e8b4c948_1200x624.png 424w, https://substackcdn.com/image/fetch/$s_!_MeP!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb835ac0-d1f7-4e4a-a5e3-bab8e8b4c948_1200x624.png 848w, https://substackcdn.com/image/fetch/$s_!_MeP!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb835ac0-d1f7-4e4a-a5e3-bab8e8b4c948_1200x624.png 1272w, https://substackcdn.com/image/fetch/$s_!_MeP!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fdb835ac0-d1f7-4e4a-a5e3-bab8e8b4c948_1200x624.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2><strong>Layer 1: Operational metrics (the efficiency floor)</strong></h2><p>These are the metrics every organization starts with, and they are necessary.</p><ul><li><p>Tasks completed per agent per hour</p></li><li><p>Cycle time reduction versus the baseline measured before deployment</p></li><li><p>Cost per transaction with and without agent involvement</p></li><li><p>Volume handled without human escalation</p></li><li><p>Error and exception rate</p></li></ul><p>Layer 1 answers the question: is the agent doing the work? For any deployment to justify itself, these numbers need to be positive. Organizations seeing 60&#8211;80% reductions in cycle time are tracking Layer 1 metrics well.</p><p>The failure mode is stopping here. Layer 1 tells you the agent is running. It does not tell you the agent is running well.</p><h2><strong>Layer 2: Decisional metrics (the quality layer)</strong></h2><p>This is where most organizations have an active measurement gap, and where the most important diagnostic information lives.</p><p>Agents make decisions. Those decisions have quality. Quality can be measured.</p><p><strong>Human override rate.</strong> When a human reviews an agent&#8217;s output, how often do they change it? A high override rate on routine tasks signals miscalibration. A low override rate on genuinely complex tasks signals overconfidence &#8212; often more dangerous than the first problem.</p><p><strong>Confidence threshold distribution.</strong> Well-designed agents signal uncertainty. Track how often your agent is operating at high, medium, and low confidence, and whether those self-assessments correlate with actual accuracy. An agent that reports high confidence but triggers frequent corrections needs retraining or rescoping.</p><p><strong>Exception escalation precision.</strong> When the agent escalates to a human, is the escalation justified? Track both the rate and the appropriateness. Agents that over-escalate are expensive. Agents that under-escalate are dangerous.</p><p><strong>Decision reversibility lag.</strong> How often is an agent decision reversed after the fact, and how much time passes before the reversal? Irreversible decisions made incorrectly compound before they surface. This metric is particularly important in financial, compliance, and customer-facing processes.</p><p><strong>Novel situation handling rate.</strong> What percentage of tasks fall outside the agent&#8217;s training distribution? This tells you something important about whether the deployment scope is well-matched to the agent&#8217;s actual capabilities.</p><p>These metrics require intentional instrumentation. They will not appear in your process management tool by default. Building the logging and evaluation layer to capture them is non-trivial work &#8212; and it is exactly the work most organizations skip because Layer 1 numbers look acceptable.</p><h2><strong>Layer 3: Adaptive capacity metrics (the transformation ceiling)</strong></h2><p>This layer measures something qualitatively different from the other two. Not whether the agent is running, and not how well it&#8217;s making decisions within the existing process. Instead: whether the process itself is expanding because the agent exists.</p><p>This is the measurement of transformational value. It is also the hardest to quantify, because it requires a counterfactual. You are measuring what is now possible that was not possible before.</p><p><strong>New capability acquisition rate.</strong> How quickly can you extend the agent to handle adjacent task types? An agent that required six weeks of development to add a new task type in its first quarter, but only two weeks by its fourth quarter, is compounding capability. One that remains at six weeks is not.</p><p><strong>Human attention quality shift.</strong> Are the humans who work alongside the agent spending more of their time on genuinely high-judgment work? Track what your people are actually doing now versus what they did before. If agent deployment simply freed them up for more of the same work, Layer 3 value is not materializing. If it redirected their attention toward decisions that actually require human judgment, it is.</p><p><strong>Process boundary expansion.</strong> Has the agent enabled the organization to take on scope that would have been infeasible before? Agentic AI&#8217;s most significant impact in mature deployments is not doing the same process faster. It is doing a different, more ambitious version of the process that was previously impractical at scale.</p><p><strong>Time-to-value on new process introductions.</strong> As you add new processes to your agent environment, how long does the ramp from introduction to operational stability take? Organizations where this number is declining have built genuine organizational capability. Those where it stays flat are running deployments, not building systems.</p><h2><strong>What this looks like in practice</strong></h2><p>Consider a representative scenario: a global financial services firm deploys an agent to handle initial client inquiry triage. Six months in, Layer 1 metrics look excellent. Inquiry cycle time is down 65%. Volume handled without escalation is up significantly.</p><p>But Layer 2 reveals a problem the leadership team hadn&#8217;t seen. The human override rate on medium-complexity inquiries is 34% &#8212; far above the 10&#8211;12% the team had assumed. And the override rate is higher on cases the agent rates as high confidence than on cases it flags as uncertain. The agent is most wrong when it thinks it&#8217;s most right.</p><p>Without Layer 2 metrics, this organization would have declared the deployment a success. With them, they have a clear retraining target, a scope adjustment to consider, and a monitoring requirement to build.</p><p>Layer 3 metrics tell a different story. The same organization discovers that because agent triage is now handling volume that previously required four full-time analysts, those analysts are available to work on relationship-intensive activities the team never had capacity for before. A new capability has emerged. That value was always latent in the process. The agent made it accessible.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!BBbb!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc881ffe7-f8e3-4825-8760-62d38faac687_1200x560.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!BBbb!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc881ffe7-f8e3-4825-8760-62d38faac687_1200x560.png 424w, https://substackcdn.com/image/fetch/$s_!BBbb!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc881ffe7-f8e3-4825-8760-62d38faac687_1200x560.png 848w, https://substackcdn.com/image/fetch/$s_!BBbb!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc881ffe7-f8e3-4825-8760-62d38faac687_1200x560.png 1272w, https://substackcdn.com/image/fetch/$s_!BBbb!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc881ffe7-f8e3-4825-8760-62d38faac687_1200x560.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!BBbb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc881ffe7-f8e3-4825-8760-62d38faac687_1200x560.png" width="1200" height="560" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c881ffe7-f8e3-4825-8760-62d38faac687_1200x560.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:560,&quot;width&quot;:1200,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:53132,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://matthewkruczek.substack.com/i/191470290?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc881ffe7-f8e3-4825-8760-62d38faac687_1200x560.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!BBbb!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc881ffe7-f8e3-4825-8760-62d38faac687_1200x560.png 424w, https://substackcdn.com/image/fetch/$s_!BBbb!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc881ffe7-f8e3-4825-8760-62d38faac687_1200x560.png 848w, https://substackcdn.com/image/fetch/$s_!BBbb!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc881ffe7-f8e3-4825-8760-62d38faac687_1200x560.png 1272w, https://substackcdn.com/image/fetch/$s_!BBbb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc881ffe7-f8e3-4825-8760-62d38faac687_1200x560.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><h2><strong>What to do this week</strong></h2><p><strong>Audit your current measurement approach.</strong> Which layer are you sitting in? If your entire agentic AI measurement program is Layer 1 metrics, you have a blind spot problem regardless of how good the numbers look.</p><p><strong>Build the Layer 2 instrumentation.</strong> The human override rate, escalation precision, and confidence calibration metrics require logging at the agent decision level. If your current deployment does not produce this data, that is your first engineering priority.</p><p><strong>Define your Layer 3 baseline.</strong> Before you can measure what new capabilities the agent creates, you need a documented picture of what the process could and could not do before deployment. This does not require sophisticated tooling. It requires a clear-eyed audit of process scope and capacity constraints.</p><p><strong>Don&#8217;t wait for the framework to be complete before sharing it.</strong> Your leadership team is asking this question now. A two-slide summary of the three-layer framework, paired with honest assessment of which layer you are currently measuring and which you are not, is more useful than a fully instrumented measurement system that arrives in six months.</p><p>The organizations that will be ahead on agentic AI measurement are not the ones with the most sophisticated dashboards. They are the ones that correctly understood what they were measuring in the first place.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://matthewkruczek.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Context Engineering! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Your Engineering Team Is Speaking a New Language: An Executive's Plain-English Guide to AI Development]]></title><description><![CDATA[If you only have a minute, here&#8217;s what you need to know.]]></description><link>https://matthewkruczek.substack.com/p/your-engineering-team-is-speaking</link><guid isPermaLink="false">https://matthewkruczek.substack.com/p/your-engineering-team-is-speaking</guid><dc:creator><![CDATA[Matthew Kruczek]]></dc:creator><pubDate>Tue, 17 Mar 2026 18:29:44 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!2uef!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe21b6f02-af7e-46c8-904b-ea9b46e36d31_2100x1107.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>If you only have a minute, here&#8217;s what you need to know.</em></p><ul><li><p>AI development has its own vocabulary, and most of it was never designed for business leaders. If you can&#8217;t follow the conversation, you can&#8217;t make informed investment decisions.</p></li><li><p>This article translates 24 essential AI terms into plain English, organized by how they relate to business decisions you&#8217;re already making.</p></li><li><p>You don&#8217;t need to become technical. You need to understand enough to ask the right questions and recognize when a vendor pitch is substance versus hand-waving.</p></li><li><p>Bookmark the glossary. Share it with your leadership team. Refer to it before your next AI briefing.</p></li></ul><p>I was explaining our AI development approach to a senior executive last week when he stopped me mid-sentence. &#8220;I don&#8217;t know what an MD file is,&#8221; he said. &#8220;What&#8217;s a skill? When you say &#8216;context,&#8217; what are you talking about?&#8221;</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://matthewkruczek.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Context Engineering! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>He wasn&#8217;t being difficult. He was being honest. And that honesty revealed a problem hiding in plain sight across every enterprise AI initiative I&#8217;ve been part of.</p><p>The people making multi-million-dollar decisions about AI strategy often can&#8217;t follow the conversations where those strategies get shaped. Not because they lack intelligence. Because the AI industry has built its own dialect, and nobody wrote a translation guide.</p><p>This is that guide.</p><h2>The vocabulary gap is a strategy gap</h2><p>A 2025 survey by Lucidworks found that 72% of C-suite executives feel confident about their organization&#8217;s AI strategy. But when researchers tested those same executives on basic AI concepts, the results told a different story. Most couldn&#8217;t accurately explain how the tools they&#8217;d approved actually work.</p><p>This isn&#8217;t an academic problem. When you can&#8217;t distinguish between a large language model and a fine-tuned model, you can&#8217;t evaluate whether your team&#8217;s proposed approach is sound. When you don&#8217;t know what a context window is, you can&#8217;t understand why your AI assistant sometimes forgets what you told it five minutes ago. When &#8220;agent&#8221; means nothing more specific than &#8220;smart bot,&#8221; you&#8217;ll overpay for simple automation dressed up in agentic packaging.</p><p>The vocabulary gap isn&#8217;t about buzzwords. It&#8217;s about decision quality.</p><h2>How AI-assisted development actually works</h2><p>Before we get to the glossary, here&#8217;s the 60-second version of how AI fits into software development today.</p><p>Your engineering teams used to write every line of code by hand. Then autocomplete tools started suggesting the next few characters. Now, AI can write entire functions, review code for security problems, generate test cases, and even architect solutions based on a description of what you want built.</p><p>Think of it like the evolution of manufacturing. Artisans making everything by hand gave way to power tools, then assembly lines, then robotic automation. The humans didn&#8217;t disappear. Their role shifted from manual labor to design, oversight, and quality control.</p><p>That&#8217;s exactly what&#8217;s happening in software development. Your best engineers are becoming orchestrators who direct AI tools rather than writing every line themselves. The ones who adapt are delivering work 60-80% faster. The ones who don&#8217;t are writing code the same way they did five years ago.</p><p>Your job as a leader isn&#8217;t to understand the technical details. It&#8217;s to understand enough to recognize where your organization sits on that spectrum, and what it takes to move forward.</p><h2>The glossary: 24 terms you&#8217;ll actually hear</h2><p>I&#8217;ve organized these into four categories that map to how you&#8217;ll encounter them in practice: the foundational concepts, the development terms, the enterprise and strategy layer, and the quality and safety terms.</p><h3><strong>The foundation: How AI thinks</strong></h3><p><strong>Artificial Intelligence (AI)</strong></p><p>Software that performs tasks typically requiring human judgment. In business context today, this almost always means software that understands and generates language, images, or code. When your teams say &#8220;AI,&#8221; they usually mean the next term.</p><p><strong>Large Language Model (LLM)</strong></p><p>The engine behind ChatGPT, Claude, and Copilot. An LLM is a program trained on enormous amounts of text that can understand questions, generate responses, write code, and analyze documents. Think of it as a very well-read colleague who has studied millions of books and conversations but has never actually worked at your company.</p><p><strong>Generative AI (GenAI)</strong></p><p>AI that creates new content: text, code, images, presentations. This is the umbrella term for the current wave of AI tools. Every time someone asks ChatGPT to draft an email, they&#8217;re using generative AI.</p><p><strong>Token</strong></p><p>The unit of measurement for AI text processing. Roughly one token equals one word, though some words get split into multiple tokens. Why this matters to you: AI services charge by token count, and every model has a maximum number of tokens it can handle at once. When your team talks about cost optimization, they&#8217;re often talking about reducing token usage.</p><p><strong>Context window</strong></p><p>The amount of information an AI can hold in its &#8220;working memory&#8221; during a single conversation. Think of it as the AI&#8217;s desk. A bigger desk lets it spread out more documents and reference more material. When someone says a model has a &#8220;200K context window,&#8221; they mean it can work with roughly 200,000 words at once. When the desk fills up, older information falls off the edge.</p><p><strong>Prompt</strong></p><p>The instruction you give an AI. &#8220;Write me a summary of this report&#8221; is a prompt. &#8220;Review this code for security issues&#8221; is a prompt. The quality of the output depends heavily on the quality of the prompt, which is why &#8220;prompt engineering&#8221; became a discipline.</p><p><strong>Inference</strong></p><p>When an AI generates a response, that process is called inference. This is the operational cost of running AI. Model training happens once (and costs millions). Inference happens every time someone asks the model a question (and costs fractions of a cent). Your cloud bills for AI are mostly inference costs.</p><p><strong>Hallucination</strong></p><p>When an AI generates information that sounds correct but is factually wrong. The model isn&#8217;t lying. It&#8217;s filling in gaps with statistically plausible text. This is why human review remains essential for anything consequential. If someone tells you their AI &#8220;never hallucinates,&#8221; they&#8217;re either confused or selling you something.</p><h3><strong>The development layer: How teams build with AI</strong></h3><p><strong>Copilot</strong></p><p>Microsoft&#8217;s branding for AI assistants embedded in their products (GitHub Copilot for code, Microsoft 365 Copilot for office work). The term has become semi-generic, the way &#8220;Kleenex&#8221; stands for tissue. When your teams say &#8220;copilot,&#8221; they mean an AI assistant that works alongside them inside the tools they already use.</p><p><strong>Agent / AI Agent</strong></p><p>An AI that can take actions on its own, not just answer questions. A chatbot waits for you to ask something. An agent can browse the web, call APIs, modify files, run code, and chain multiple steps together to complete a task. An agent is the difference between a librarian who finds books for you and an assistant who researches, summarizes, and drafts the report.</p><p><strong>Multi-agent</strong></p><p>Multiple specialized AI agents working together on a task. One agent handles requirements analysis. Another writes code. A third reviews it for security issues. A fourth runs tests. This mirrors how human teams divide labor, except agents can run in parallel and hand off work without scheduling meetings.</p><p><strong>Skill</strong></p><p>A packaged set of instructions that teaches an AI agent how to perform a specific task the way your organization does it. Without skills, an agent writes code in a generic style. With your company&#8217;s skill loaded, it follows your naming conventions, uses your preferred frameworks, and applies your security standards. Skills are how you turn a general-purpose AI into one that understands &#8220;how we do things here.&#8221;</p><p><strong>Orchestration</strong></p><p>The act of coordinating multiple AI agents, tools, and workflows to accomplish a complex task. Your senior engineers are increasingly becoming orchestrators rather than hands-on coders. They design the workflow, assign agents to each step, set quality checkpoints, and validate the results.</p><p><strong>Markdown</strong></p><p>A simple text formatting language that uses symbols instead of toolbar buttons. A hashtag (#) makes a heading. Asterisks make text bold. Dashes create bullet lists. AI tools use markdown extensively because it&#8217;s lightweight and readable by both humans and machines. When your team mentions markdown, they&#8217;re talking about a plain text format, not a programming language.</p><p><strong>MCP (Model Context Protocol)</strong></p><p>An open standard that lets AI agents connect to external tools and data sources. Think of it like USB for AI. Before USB, every peripheral needed its own proprietary cable. MCP does the same for AI integrations: one standard protocol that connects any AI model to any tool, database, or service.</p><p><strong>RAG (Retrieval-Augmented Generation)</strong></p><p>A technique that lets AI pull in relevant information from your company&#8217;s documents before generating a response. Instead of relying only on what the model learned during training, RAG searches your knowledge bases in real time. This is how you make AI answers specific to your organization without retraining the entire model.</p><p><strong>Fine-tuning</strong></p><p>Retraining an AI model on your company&#8217;s specific data so it performs better on your use cases. This is more expensive and complex than RAG but produces a model that deeply understands your domain. Most organizations start with RAG and fine-tune only when they&#8217;ve proven the use case justifies the investment.</p><h3><strong>The strategy layer: Enterprise AI decisions</strong></h3><p><strong>Prompt engineering</strong></p><p>The discipline of crafting effective instructions for AI. Early AI adoption treated this as an art. It&#8217;s now becoming a systematic practice with documented patterns and measurable outcomes. Good prompt engineering is the difference between AI that produces generic output and AI that delivers exactly what you need.</p><p><strong>Context engineering</strong></p><p>The practice of controlling everything an AI can see when it processes a request: the prompt, the documents, the conversation history, the organizational knowledge. If prompt engineering is writing a good question, context engineering is preparing the entire briefing package. This is where the real enterprise value lives.</p><p><strong>Specification (in agent development)</strong></p><p>A document that describes what success looks like rather than step-by-step instructions. Traditional requirements tell developers how to build something. Specifications for agents describe the desired outcome, constraints, and quality criteria, then let the agent determine the best approach. Writing good specifications is becoming one of the most valuable skills in engineering.</p><p><strong>Agentic AI</strong></p><p>AI systems designed to act autonomously toward goals. Rather than answering individual questions, agentic AI breaks down complex objectives, creates plans, executes steps, evaluates results, and adjusts course. This is the frontier of enterprise AI, where the technology moves from assistant to operator.</p><h3><strong>The safety layer: Quality and risk</strong></h3><p><strong>Guardrails</strong></p><p>Rules and constraints that prevent AI from doing things it shouldn&#8217;t. Content filters, spending limits, approval gates before actions are taken, restrictions on what data the AI can access. Guardrails are how you maintain control while giving AI enough autonomy to be useful. More guardrails means more safety but less speed. Finding the right balance is a leadership decision.</p><p><strong>Grounding</strong></p><p>Connecting AI outputs to verified information sources. An ungrounded response comes purely from the model&#8217;s training data (which may be outdated or wrong). A grounded response cites specific, verifiable sources. Grounding is how you reduce hallucinations and make AI outputs trustworthy enough for business decisions.</p><p><strong>Human-in-the-loop</strong></p><p>A workflow design where humans review and approve AI decisions at critical points. Not everything needs human approval (an AI can auto-format code without asking). But security changes, financial transactions, and customer-facing content should have a human checkpoint. The question isn&#8217;t whether to include humans but where in the process they add the most value.</p><h2>Putting it all together: A real-world scenario</h2><p>The glossary gives you definitions. But the real value is understanding how these pieces connect. Let me walk you through a scenario your teams are probably living right now.</p><p>Imagine your engineering team gets a request: build a new customer portal that lets clients track their order status, contact support, and manage their account. Here&#8217;s how AI-assisted development handles this, with every term from the glossary showing up naturally.</p><p><strong>The starting point.</strong> A senior engineer sits down with an <strong>AI agent</strong>, not a simple chatbot, but one that can take actions, read files, and execute multi-step work. The engineer writes a <strong>prompt</strong>: &#8220;Build a customer portal with order tracking, support chat, and account management. Follow our company&#8217;s security and design standards.&#8221;</p><p>That prompt is short. But here&#8217;s where <strong>context engineering</strong> comes in. The agent doesn&#8217;t just see those two sentences. It also loads the team&#8217;s <strong>skills</strong>, pre-packaged instructions that encode how this company builds software. One skill says &#8220;we use React for frontends.&#8221; Another says &#8220;all customer data must be encrypted at rest and accessed through our API gateway.&#8221; A third contains the company&#8217;s design system with approved colors, fonts, and component patterns.</p><p>All of that, the prompt, the skills, the conversation history, the referenced documents, fills the agent&#8217;s <strong>context window</strong>. Think of it this way: if the context window is a desk, the prompt is the sticky note with today&#8217;s assignment. The skills are the company handbook, the style guide, and the security policy. The referenced codebase documentation is the stack of technical drawings. Everything has to fit on that desk at the same time, and the desk has a fixed size.</p><p>This is why <strong>tokens</strong> matter to your budget. Every word in those skills, every line of referenced documentation, every back-and-forth message in the conversation consumes tokens. A skill that&#8217;s 3,000 tokens is taking up desk space that could be used for the actual work. Multiply that across dozens of agents running hundreds of requests per day, and token efficiency directly affects your cloud spend. A bloated set of instructions that consumes 50,000 tokens per request costs real money at scale, and it crowds out the space the agent needs to reason about the actual problem.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!2uef!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe21b6f02-af7e-46c8-904b-ea9b46e36d31_2100x1107.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!2uef!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe21b6f02-af7e-46c8-904b-ea9b46e36d31_2100x1107.png 424w, https://substackcdn.com/image/fetch/$s_!2uef!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe21b6f02-af7e-46c8-904b-ea9b46e36d31_2100x1107.png 848w, https://substackcdn.com/image/fetch/$s_!2uef!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe21b6f02-af7e-46c8-904b-ea9b46e36d31_2100x1107.png 1272w, https://substackcdn.com/image/fetch/$s_!2uef!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe21b6f02-af7e-46c8-904b-ea9b46e36d31_2100x1107.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!2uef!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe21b6f02-af7e-46c8-904b-ea9b46e36d31_2100x1107.png" width="1456" height="768" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/e21b6f02-af7e-46c8-904b-ea9b46e36d31_2100x1107.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:768,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:201445,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://matthewkruczek.substack.com/i/191283077?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe21b6f02-af7e-46c8-904b-ea9b46e36d31_2100x1107.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!2uef!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe21b6f02-af7e-46c8-904b-ea9b46e36d31_2100x1107.png 424w, https://substackcdn.com/image/fetch/$s_!2uef!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe21b6f02-af7e-46c8-904b-ea9b46e36d31_2100x1107.png 848w, https://substackcdn.com/image/fetch/$s_!2uef!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe21b6f02-af7e-46c8-904b-ea9b46e36d31_2100x1107.png 1272w, https://substackcdn.com/image/fetch/$s_!2uef!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe21b6f02-af7e-46c8-904b-ea9b46e36d31_2100x1107.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>Where context limits hit.</strong> Here&#8217;s the moment that catches most teams off guard. The engineer asks the agent to review the entire existing codebase (150,000 lines) alongside the new portal requirements. But the agent&#8217;s context window is 200,000 tokens. The codebase alone consumes 120,000 tokens. The skills take another 15,000. The conversation history is 10,000. That leaves just 55,000 tokens for the agent to actually think and generate code. Performance drops. The agent starts &#8220;forgetting&#8221; instructions from earlier in the conversation because they&#8217;ve fallen off the edge of the desk.</p><p>This is why your team talks about <strong>RAG</strong> (retrieval-augmented generation). Instead of cramming the entire codebase into the context window, RAG lets the agent search for just the relevant files when it needs them. The agent pulls in only the authentication module when working on login, only the order database schema when building the tracking page. The desk stays clean. The agent stays sharp.</p><p><strong>The multi-agent workflow.</strong> Now the work splits across specialized <strong>agents</strong> in a <strong>multi-agent</strong> setup. One agent analyzes the requirements and produces a technical <strong>specification</strong>, a document describing what success looks like rather than step-by-step instructions. A second agent takes that specification and generates the code. A third agent reviews the code for security vulnerabilities. A fourth writes and runs tests.</p><p>Each agent has its own context window, its own desk. The security review agent loads the company&#8217;s security <strong>skill</strong> (which includes OWASP compliance requirements and the company&#8217;s specific data handling policies). It doesn&#8217;t need the design system skill. That keeps its desk clear for focused security analysis.</p><p><strong>Orchestration</strong> is the act of coordinating all of this. The senior engineer isn&#8217;t writing code line by line. They&#8217;re designing the workflow: which agent handles which task, what quality gates sit between each step, where a <strong>human-in-the-loop</strong> checkpoint is needed. Security changes? Human approval required. Color adjustments to a button? The agent handles it autonomously.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!QWQg!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29ad1d42-8558-49fa-8754-c936c05cc17d_2973x804.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!QWQg!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29ad1d42-8558-49fa-8754-c936c05cc17d_2973x804.png 424w, https://substackcdn.com/image/fetch/$s_!QWQg!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29ad1d42-8558-49fa-8754-c936c05cc17d_2973x804.png 848w, https://substackcdn.com/image/fetch/$s_!QWQg!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29ad1d42-8558-49fa-8754-c936c05cc17d_2973x804.png 1272w, https://substackcdn.com/image/fetch/$s_!QWQg!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29ad1d42-8558-49fa-8754-c936c05cc17d_2973x804.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!QWQg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29ad1d42-8558-49fa-8754-c936c05cc17d_2973x804.png" width="1456" height="394" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/29ad1d42-8558-49fa-8754-c936c05cc17d_2973x804.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:394,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:207235,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://matthewkruczek.substack.com/i/191283077?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29ad1d42-8558-49fa-8754-c936c05cc17d_2973x804.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!QWQg!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29ad1d42-8558-49fa-8754-c936c05cc17d_2973x804.png 424w, https://substackcdn.com/image/fetch/$s_!QWQg!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29ad1d42-8558-49fa-8754-c936c05cc17d_2973x804.png 848w, https://substackcdn.com/image/fetch/$s_!QWQg!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29ad1d42-8558-49fa-8754-c936c05cc17d_2973x804.png 1272w, https://substackcdn.com/image/fetch/$s_!QWQg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F29ad1d42-8558-49fa-8754-c936c05cc17d_2973x804.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>The connection layer.</strong> The portal needs to pull order data from the existing ERP system and send support tickets to the CRM. This is where <strong>MCP</strong> (Model Context Protocol) comes in. Rather than building custom integrations for each AI tool, MCP provides a standard interface. The agent connects to the ERP through an MCP server the same way it connects to the CRM, the same way it connects to the company&#8217;s internal documentation. One protocol, many connections, like USB for AI.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!KYwH!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc29e04d-a8d1-4762-8878-88ff4ac59a77_1740x1899.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!KYwH!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc29e04d-a8d1-4762-8878-88ff4ac59a77_1740x1899.png 424w, https://substackcdn.com/image/fetch/$s_!KYwH!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc29e04d-a8d1-4762-8878-88ff4ac59a77_1740x1899.png 848w, https://substackcdn.com/image/fetch/$s_!KYwH!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc29e04d-a8d1-4762-8878-88ff4ac59a77_1740x1899.png 1272w, https://substackcdn.com/image/fetch/$s_!KYwH!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc29e04d-a8d1-4762-8878-88ff4ac59a77_1740x1899.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!KYwH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc29e04d-a8d1-4762-8878-88ff4ac59a77_1740x1899.png" width="1456" height="1589" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/bc29e04d-a8d1-4762-8878-88ff4ac59a77_1740x1899.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1589,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:287947,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://matthewkruczek.substack.com/i/191283077?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc29e04d-a8d1-4762-8878-88ff4ac59a77_1740x1899.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!KYwH!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc29e04d-a8d1-4762-8878-88ff4ac59a77_1740x1899.png 424w, https://substackcdn.com/image/fetch/$s_!KYwH!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc29e04d-a8d1-4762-8878-88ff4ac59a77_1740x1899.png 848w, https://substackcdn.com/image/fetch/$s_!KYwH!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc29e04d-a8d1-4762-8878-88ff4ac59a77_1740x1899.png 1272w, https://substackcdn.com/image/fetch/$s_!KYwH!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fbc29e04d-a8d1-4762-8878-88ff4ac59a77_1740x1899.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong>Quality at the end.</strong> Before anything ships, <strong>guardrails</strong> catch problems. The agent can&#8217;t deploy directly to production without approval. It can&#8217;t access customer financial data outside the approved API. <strong>Grounding</strong> ensures that when the agent generates documentation for the portal, it cites actual API endpoints and real database fields rather than <strong>hallucinating</strong> plausible-sounding but nonexistent ones.</p><p><strong>What the executive sees.</strong> From your chair, here&#8217;s what happened: a feature that used to take a team of five developers three months was delivered in three weeks by two engineers orchestrating AI agents. The code follows your company&#8217;s standards (because those standards were encoded as skills). The security review was thorough (because a specialized agent ran checks against your actual compliance requirements). And the cost was measurable in tokens consumed, agent time used, and human hours for oversight.</p><p>That&#8217;s the full picture. Every term in the glossary has a role. None of them exist in isolation. And when your team throws these words around in their next briefing, you&#8217;ll know exactly how the pieces connect.</p><h2>How to use this in your next meeting</h2><p>You don&#8217;t need to memorize these definitions. You need to know enough to ask three questions:</p><p><strong>&#8220;What&#8217;s the context window for our current setup, and are we hitting limits?&#8221;</strong> This tells you whether your teams are constrained by the AI&#8217;s working memory, which directly affects output quality.</p><p><strong>&#8220;Where are the human-in-the-loop checkpoints, and why there?&#8221;</strong> This reveals how much autonomy the AI has and whether the risk controls match your tolerance.</p><p><strong>&#8220;What skills have we built, and what institutional knowledge are we still missing?&#8221;</strong> This tells you whether your AI investment is accumulating organizational value or starting from scratch every time.</p><p>Those three questions will give you more strategic insight than any vendor demo.</p><h2>The real risk isn&#8217;t the technology</h2><p>Every technology wave comes with its own vocabulary. Cloud computing brought us &#8220;elastic scaling&#8221; and &#8220;microservices.&#8221; Mobile brought us &#8220;responsive design&#8221; and &#8220;push notifications.&#8221; Leaders figured those out because they had to.</p><p>AI&#8217;s vocabulary wave is bigger, faster, and more consequential. The decisions being made right now, about agents, context engineering, multi-agent orchestration, and skill development, will shape your technology organization for the next decade.</p><p>You don&#8217;t need to write code. You don&#8217;t need to understand neural network architecture. But you do need to understand the language well enough to evaluate whether the strategy your team is proposing will actually work.</p><p>The executives who learn this vocabulary won&#8217;t just follow the conversation. They&#8217;ll lead it.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://matthewkruczek.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Context Engineering! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[Agent Harnesses Don't Need More Layers. They Need Fewer.]]></title><description><![CDATA[If you only have a minute, here's what you need to know.]]></description><link>https://matthewkruczek.substack.com/p/agent-harnesses-dont-need-more-layers</link><guid isPermaLink="false">https://matthewkruczek.substack.com/p/agent-harnesses-dont-need-more-layers</guid><dc:creator><![CDATA[Matthew Kruczek]]></dc:creator><pubDate>Mon, 16 Mar 2026 14:37:11 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!VXWr!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec4aca9e-0c9d-493e-b1ff-064c3a072af8_736x736.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>If you only have a minute, here's what you need to know.</em></p><ul><li><p>Agent harnesses, the infrastructure wrapping AI agents for production use, are becoming the defining architecture pattern of 2026. But most enterprises are building them wrong.</p></li><li><p>Evidence from Vercel, Manus, and OpenAI&#8217;s Codex shows that stripping tools and complexity consistently outperforms adding more scaffolding. Vercel cut 15 specialized tools down to 2 and saw accuracy jump from 80% to 100%.</p></li><li><p>Microsoft&#8217;s Agent Framework takes a disciplined approach: approval workflows, context compaction strategies, and dual-language support in Python and .NET, without the bloat that plagues custom harness implementations.</p></li><li><p>The enterprises that will succeed with agent infrastructure aren&#8217;t the ones with the most sophisticated orchestration layers. They&#8217;re the ones with the fewest.</p></li><li><p>If your agent harness has more abstraction layers than your agent has tools, you&#8217;ve already lost the plot.</p><p></p><div class="captioned-button-wrap" data-attrs="{&quot;url&quot;:&quot;https://matthewkruczek.substack.com/p/agent-harnesses-dont-need-more-layers?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="CaptionedButtonToDOM"><div class="preamble"><p class="cta-caption">Thanks for reading Context Engineering! This post is public so feel free to share it.</p></div><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://matthewkruczek.substack.com/p/agent-harnesses-dont-need-more-layers?utm_source=substack&utm_medium=email&utm_content=share&action=share&quot;,&quot;text&quot;:&quot;Share&quot;}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://matthewkruczek.substack.com/p/agent-harnesses-dont-need-more-layers?utm_source=substack&utm_medium=email&utm_content=share&action=share"><span>Share</span></a></p></div></li></ul><div><hr></div><p>The AI industry spent 2025 building agents. 2026 is the year we figure out how to control them.</p><p>A new term is circulating in enterprise architecture circles: <strong>agent harness</strong>. It refers to the infrastructure layer that sits between an AI model and the real world, managing tool access, approval workflows, context windows, error recovery, and state persistence. Think of the model as the engine and the harness as the car. Without the car, the engine is impressive but useless. Without a good car, the engine destroys itself.</p><p>The concept isn&#8217;t new. Anyone running agents in production has been building some version of a harness for the past year. What&#8217;s new is that the industry is now treating harness engineering as a first-class discipline, with dedicated frameworks, design patterns, and an emerging consensus that <strong>the harness, not the model, determines whether agents succeed or fail in production</strong>.</p><p>I agree with that premise. But I&#8217;m watching enterprises draw exactly the wrong conclusion from it.</p><h2><strong>The complexity trap</strong></h2><p>The natural instinct when you recognize the harness matters is to build more of it. More abstraction layers. More specialized tools. More governance checkpoints. More orchestration logic. Enterprise architects see the agent harness and think: finally, something I can over-engineer.</p><p>The evidence points in the opposite direction.</p><p>Vercel&#8217;s agent team ran an experiment that should be required reading for every enterprise architect. They had 15 specialized tools powering their AI coding agents. They removed 13 of them, keeping only two: bash execution and SQL queries. The result? Accuracy jumped from 80% to 100%. Token usage dropped 37%. Speed improved 3.5x. Fewer tools, dramatically better outcomes.</p><p>Manus, the autonomous agent framework, tells the same story. The team rebuilt their agent system four times. The greatest performance gains came not from adding capabilities but from removing complexity. They implemented filesystem-as-memory, aggressive context compaction (100:1 input-to-output ratio), and KV-cache optimization. The result was a 10x cost reduction through pure infrastructure simplification.</p><p>OpenAI&#8217;s internal Codex agents converged on identical principles independently. Minimal, general-purpose tools. External state persistence through git and files. Structured error retention. Strict context discipline.</p><p>Three separate teams, three different organizations, one conclusion: <strong>the best agent harness is the one with the least in it</strong>.</p><h2><strong>Why less works better</strong></h2><p>This isn&#8217;t counterintuitive once you understand the mechanics.</p><p>Every tool you add to an agent&#8217;s context window competes for the model&#8217;s attention. I wrote about this in my piece on <a href="https://matthewkruczek.ai/blog/progressive-disclosure-mcp-servers.html">progressive disclosure for MCP servers</a>: 400 tools can consume 400,000+ tokens, exceeding even the largest context windows. But the problem isn&#8217;t just token count. It&#8217;s decision fatigue. Models, like humans, make worse choices when presented with too many options.</p><p>Specialized tools also create routing problems. When you give an agent 15 ways to accomplish similar tasks, it spends reasoning cycles figuring out which tool to use instead of solving the actual problem. Strip it down to bash and a database connection, and the model focuses its reasoning on what matters: accomplishing the objective.</p><p><a href="http://www.incompleteideas.net/IncIdeas/BitterLesson.html">Richard Sutton&#8217;s bitter lesson</a> from machine learning applies directly here. General methods that use computation beat specialized methods that try to encode human knowledge about the domain. Complex scaffolding becomes obsolete as models improve. The harness should simplify with model upgrades, not accumulate complexity.</p><h2><strong>What Microsoft gets right</strong></h2><p>This is the lens through which I&#8217;ve been evaluating Microsoft&#8217;s Agent Framework, and its <a href="https://devblogs.microsoft.com/agent-framework/agent-harness-in-agent-framework/">recently published agent harness patterns</a> specifically.</p><p>The framework focuses on three building blocks: local shell execution with approval gates, hosted shell in managed environments, and context compaction. That&#8217;s it. Three patterns, not thirty.</p><p>The approval workflow is particularly well-designed. In Python, you decorate a tool with </p><div class="highlighted_code_block" data-attrs="{&quot;language&quot;:&quot;plaintext&quot;,&quot;nodeId&quot;:&quot;22302648-6112-430c-b515-4e099006624b&quot;}" data-component-name="HighlightedCodeBlockToDOM"><pre class="shiki"><code class="language-plaintext">@tool(approval_mode="always_require") </code></pre></div><p>and the framework handles the rest. In .NET, you wrap tools with </p><div class="highlighted_code_block" data-attrs="{&quot;language&quot;:&quot;plaintext&quot;,&quot;nodeId&quot;:&quot;6419da2e-e08b-40e4-9655-87a1e45a63a6&quot;}" data-component-name="HighlightedCodeBlockToDOM"><pre class="shiki"><code class="language-plaintext">ApprovalRequiredAIFunction. </code></pre></div><p>The pattern is explicit and minimal. There&#8217;s no sprawling governance layer, just a clear gate at the point where the agent wants to do something irreversible.</p><p>Context compaction is where the real discipline shows. Long-running agent sessions inevitably exceed context windows. Microsoft&#8217;s approach offers composable strategies: sliding window, tool result compaction, and truncation, combined through a pipeline. The .NET implementation chains</p><div class="highlighted_code_block" data-attrs="{&quot;language&quot;:&quot;plaintext&quot;,&quot;nodeId&quot;:&quot;0341b6c7-e701-48f7-be4e-99c2e4201d15&quot;}" data-component-name="HighlightedCodeBlockToDOM"><pre class="shiki"><code class="language-plaintext">ToolResultCompactionStrategy, SlidingWindowCompactionStrategy, </code></pre></div><p>and </p><div class="highlighted_code_block" data-attrs="{&quot;language&quot;:&quot;plaintext&quot;,&quot;nodeId&quot;:&quot;953f280f-1840-41f4-8042-7be4934defdc&quot;}" data-component-name="HighlightedCodeBlockToDOM"><pre class="shiki"><code class="language-plaintext">TruncationCompactionStrategy </code></pre></div><p>into a single </p><div class="highlighted_code_block" data-attrs="{&quot;language&quot;:&quot;plaintext&quot;,&quot;nodeId&quot;:&quot;3661a9b3-8bf3-4f70-aa07-0c42d2f04ef4&quot;}" data-component-name="HighlightedCodeBlockToDOM"><pre class="shiki"><code class="language-plaintext">PipelineCompactionStrategy. </code></pre></div><p>Configurable. Composable. Not over-abstracted.</p><p>And the dual-language support in Python and .NET matters more than it might seem. Enterprise teams aren&#8217;t monolingual. The same harness patterns working identically in both ecosystems means your .NET backend team and your Python data science team can build agent infrastructure using shared concepts. That&#8217;s practical enterprise architecture, not marketing.</p><h2><strong>The governance question</strong></h2><p>I can already hear the objection: &#8220;But we need governance. Compliance. Audit trails. We can&#8217;t just give agents bash access and hope for the best.&#8221;</p><p>Fair. But governance doesn&#8217;t require complexity. The <a href="https://www.cncf.io/blog/2026/01/23/the-autonomous-enterprise-and-the-four-pillars-of-platform-control-2026-forecast/">CNCF&#8217;s framework for autonomous enterprise governance</a> identifies four pillars: golden paths (pre-approved configurations), guardrails (policy enforcement), safety nets (automated recovery), and manual review gates. Notice what&#8217;s missing: there&#8217;s no pillar for &#8220;add seventeen orchestration layers.&#8221;</p><p>The most effective governance I&#8217;ve seen in production agent systems follows a simple rule: <strong>intervene only when the model can&#8217;t self-correct</strong>. That means approval gates for irreversible actions, sandboxing for execution environments, audit logging for everything, and nothing else. Every additional governance mechanism is a tax on agent performance that needs to justify its existence with a specific risk it mitigates.</p><h2><strong>What to do this week</strong></h2><p>If you&#8217;re building or evaluating agent harness infrastructure, here&#8217;s my recommendation:</p><p><strong>Audit your tool count.</strong> If your agents have access to more than 5-7 tools, run the Vercel experiment. Strip down to the minimum general-purpose set and measure the difference. You may be surprised.</p><p><strong>Adopt a framework, don&#8217;t build from scratch.</strong> Microsoft&#8217;s Agent Framework, LangGraph, or similar production-tested frameworks have already solved the foundational problems. Your engineering effort should go into your specific approval workflows and domain logic, not reinventing context management.</p><p><strong>Measure harness complexity as a cost.</strong> Every abstraction layer, every custom tool, every governance checkpoint has a performance cost in tokens, latency, and error surface. Track it. If a layer doesn&#8217;t measurably improve outcomes, remove it.</p><p><strong>Design for deletion.</strong> As models improve, your harness should get simpler, not more complex. Build infrastructure that&#8217;s easy to remove. The scaffolding you need today for GPT-4o may be unnecessary for whatever ships next quarter.</p><p>The enterprises that will win the agent infrastructure race aren&#8217;t building the most sophisticated harnesses. They&#8217;re building the most disciplined ones. And discipline, in this context, means knowing what to leave out.</p><div><hr></div><h2><strong>References</strong></h2><ol><li><p>Microsoft Agent Framework. &#8220;Agent Harness in Agent Framework.&#8221; March 12, 2026. <a href="https://devblogs.microsoft.com/agent-framework/agent-harness-in-agent-framework/">https://devblogs.microsoft.com/agent-framework/agent-harness-in-agent-framework/</a></p></li><li><p>Pappas, E. &#8220;The Agent Harness Is the Architecture (and Your Model Is Not the Bottleneck).&#8221; DEV Community, 2026. <a href="https://dev.to/epappas/the-agent-harness-is-the-architecture-and-your-model-is-not-the-bottleneck-3bjd">https://dev.to/epappas/the-agent-harness-is-the-architecture-and-your-model-is-not-the-bottleneck-3bjd</a></p></li><li><p>HTEKDev. &#8220;Agent Harnesses: Why 2026 Isn&#8217;t About More Agents, It&#8217;s About Controlling Them.&#8221; DEV Community, 2026. <a href="https://dev.to/htekdev/agent-harnesses-why-2026-isnt-about-more-agents-its-about-controlling-them-1f24">https://dev.to/htekdev/agent-harnesses-why-2026-isnt-about-more-agents-its-about-controlling-them-1f24</a></p></li><li><p>OpenAI. &#8220;Harness Engineering: Codex Agents Power Large-Scale Software Development.&#8221; InfoQ, February 2026. <a href="https://www.infoq.com/news/2026/02/openai-harness-engineering-codex/">https://www.infoq.com/news/2026/02/openai-harness-engineering-codex/</a></p></li><li><p>Gupta, A. &#8220;2025 Was Agents. 2026 Is Agent Harnesses.&#8221; Medium, 2026. <a href="https://aakashgupta.medium.com/2025-was-agents-2026-is-agent-harnesses-heres-why-that-changes-everything-073e9877655e">https://aakashgupta.medium.com/2025-was-agents-2026-is-agent-harnesses-heres-why-that-changes-everything-073e9877655e</a></p></li><li><p>Kruczek, M. &#8220;Progressive Disclosure for MCP Servers: A Design Pattern for Scalable AI Tool Integration.&#8221; <a href="http://matthewkruczek.ai/">matthewkruczek.ai</a>.</p></li><li><p>CNCF. &#8220;The Autonomous Enterprise and the Four Pillars of Platform Control: 2026 Forecast.&#8221; January 23, 2026. <a href="https://www.cncf.io/blog/2026/01/23/the-autonomous-enterprise-and-the-four-pillars-of-platform-control-2026-forecast/">https://www.cncf.io/blog/2026/01/23/the-autonomous-enterprise-and-the-four-pillars-of-platform-control-2026-forecast/</a></p></li><li><p>Sutton, R. &#8220;The Bitter Lesson.&#8221; 2019. <a href="http://www.incompleteideas.net/IncIdeas/BitterLesson.html">http://www.incompleteideas.net/IncIdeas/BitterLesson.html</a></p></li></ol><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://matthewkruczek.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Context Engineering! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item><item><title><![CDATA[Why I'm Starting This]]></title><description><![CDATA[Most AI content falls into one of two categories.]]></description><link>https://matthewkruczek.substack.com/p/coming-soon</link><guid isPermaLink="false">https://matthewkruczek.substack.com/p/coming-soon</guid><dc:creator><![CDATA[Matthew Kruczek]]></dc:creator><pubDate>Sun, 15 Mar 2026 17:17:41 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!VXWr!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fec4aca9e-0c9d-493e-b1ff-064c3a072af8_736x736.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Most AI content falls into one of two categories. </p><p>The first is academic. Well-sourced, carefully hedged, and almost entirely disconnected from what it actually takes to deploy an AI system inside a company with compliance requirements, legacy infrastructure, and people who&#8217;ve been burned by a previous technology promise. </p><p>The second is enthusiasm. Tools, demos, benchmark comparisons. Useful for keeping up with what&#8217;s shipping, not so useful for figuring out what to build and why. </p><p>Context Engineering is an attempt at a third thing: a practitioner&#8217;s record of what I&#8217;m actually seeing. </p><p>I&#8217;ve spent years deploying AI systems at Fortune 500 scale. Before that, I created technical courses that reached <strong>17 million developers</strong> on Pluralsight. I&#8217;ve been on the ground for a lot of AI projects, both the ones that shipped and the ones that didn&#8217;t.</p><p> The difference, in my experience, is almost always context.</p><p>Not model capability. Not prompt phrasing. Context: what the system knows, when it knows it, how it&#8217;s structured, and what happens to it at scale. That&#8217;s the skill nobody is talking about clearly, and it&#8217;s the skill that actually determines whether an AI project ships or stalls in the demo phase. </p><p>That&#8217;s what this newsletter covers. Every issue is one observation from the field. Written for the engineers designing the systems and for the leaders deciding whether to invest in them. Specific enough to be useful, honest enough to be trusted. </p><p>If that sounds worth reading, subscribe below and share it with one person who&#8217;d get something out of it. </p><p><strong>First issue coming this week.</strong></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://matthewkruczek.substack.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://matthewkruczek.substack.com/subscribe?"><span>Subscribe now</span></a></p><p></p>]]></content:encoded></item></channel></rss>