Where AI Creates Repeatable Value and Where It Doesn't
Most organisations can point to where AI is being used. Fewer can say where it has stabilised.
The difference is not semantic. Activity and value are easy to confuse in the early stages of adoption. Pilots expand. Tools get embedded in workflows. Teams report productivity improvements. Usage dashboards show upward curves. From a distance, this looks like progress. And sometimes it is.
But until behaviour stabilises within a role, until a function is using AI in a way that is predictable, owned, and no longer dependent on individual enthusiasm or a single champion’s energy, what you’re observing is experimentation, not value creation. The distinction matters because organisations routinely make scaling decisions on the basis of experimental signal, treating early activity as evidence that value has formed when it has only been glimpsed.
Experimentation is valuable. It generates learning, surfaces possibilities, builds familiarity. It is also inherently volatile. A team that adopted a tool last quarter may have moved on this quarter. A workflow that showed promise under one manager may not survive a reorganisation. A use case that produced impressive results in a pilot may fail to replicate when ownership transfers from the person who built it to the team expected to maintain it. Activity without stability is signal without structure.
What changes the picture is when a role begins to treat AI not as something it is testing but as something it operates through. The variance in usage drops. Ownership becomes clear. The function stops debating whether to use AI for a given task and starts debating how to improve the way it’s used. Conversations shift from “should we try this” to “how do we do this better.” This is where repeatable economic value begins, and it is visible in behaviour before it shows up in any financial metric.
The question for leaders is not whether AI is being used. It is whether that usage has crossed the threshold from experimentation into something durable enough to build on.
This is precisely what Quaie’s Q1 2026 fieldwork is designed to measure. When we ask ten executive roles about adoption stage and confidence in durable value, the hypothesis is that the results will not distribute evenly across the cohort. Some roles are likely to cluster toward experimentation. Others will have moved into limited production use. A smaller number may report scaled deployment. The more significant question is not who is furthest ahead, but how sharply confidence in durable value tracks adoption stage. This is the relationship the Role Shift Index is designed to capture, placing each role on the adoption spectrum and tracking how that position shifts over time. If the pattern holds as expected, roles at scaled deployment will report substantially higher confidence that their AI initiatives will produce lasting economic value than roles still at the experimentation stage. And that gap is likely to map to role more cleanly than to any other variable, including company size, revenue band, or the specific AI applications being used. Quaie’s Organisational Adoption Gradient, the distance between the most advanced and least advanced roles, is designed to make this divergence visible rather than allowing it to be concealed within an enterprise-level average.
This suggests something that most AI benchmarking misses entirely. Value does not emerge evenly across an organisation and then get recognised. It concentrates in specific roles first. And the roles where it concentrates are the ones whose context allows experimentation to convert into something durable: decision authority over the relevant workflow, short enough feedback loops to iterate quickly, and proximity to outcomes that can be measured without ambiguity.
Marketing and customer service are the two functional areas most commonly identified in early research as having crossed from experimentation into formal approval or dedicated budget.¹ The reasons are structural. Both areas involve high-frequency, repeatable tasks where AI can be tested against clear performance baselines. A marketing team running AI-assisted campaign optimisation can see results within days. A customer service operation using AI for triage and response can measure impact within weeks. The feedback loop is tight enough for experimentation to stabilise relatively quickly.
But even within these areas, confidence in durable value is likely to remain uneven. The crossing point from “we’re trying this” to “this is how we work now” is not a clean threshold. It is a gradual stabilisation that is easier to see in retrospect than in the moment. A function can be well into production use and still harbour uncertainty about whether the value will persist through a budget review, a leadership change, or a shift in strategic priorities.
This is why measuring adoption by what has been deployed misses the point. Deployment is a moment. Stabilisation is a process. And the process looks different depending on which role you’re observing.
Consider two roles in the same organisation. A CTO deploys AI tooling across engineering and sees rapid uptake. Technical teams are comfortable with new tools. Feedback loops are short. The CTO has direct authority over the workflow, and when something doesn’t work, the team can adjust without waiting for cross-functional approval. Within a quarter, usage has stabilised. The experimentation phase is over.
The same organisation’s CMO deploys AI in campaign planning and sees a different trajectory. Creative teams are less comfortable with AI-assisted workflows. Measurement is harder because marketing outcomes are influenced by variables outside the CMO’s control. Authority over the workflow is shared with agencies and partners who have their own views on AI. Six months in, usage is inconsistent. Some team members have adopted it. Others have reverted to previous methods. The CMO describes it as “in progress.” In practice, it is still experimental.
Both roles deployed AI. One is approaching repeatable value. The other is still in experimentation, even if nobody describes it that way internally. The Role Lead-Lag Ranking between these two roles would show a widening temporal gap, the CTO pulling further ahead while the CMO’s position remains static, a divergence invisible to any enterprise-level metric. This dynamic is also captured by the Role Influence Index, which measures the relative influence each leadership role exerts over adoption decisions. The CTO’s direct authority over tooling and workflow makes the role a primary catalyst; the CMO’s shared authority with external partners and creative functions positions it closer to validator or conditional adopter, which in part explains the slower path to stabilisation.
The practical implication is uncomfortable but important. Early wins in one function do not predict success elsewhere. What is working in marketing is working because of marketing’s specific role context. Finance has a different context, different blockers, and a different path to stabilisation. Operations has another. The CHRO faces yet another, workforce readiness questions that no other role is addressing, with feedback loops measured in quarters rather than days. The instinct to generalise (”AI is working here, so let’s accelerate it everywhere”) misreads what is actually happening. It mistakes a role-specific outcome for an organisational one.
The more useful question for any leadership team is not “where are we using AI?” but “where has AI use become predictable and owned?” Where is the variance dropping? Where has a function stopped experimenting and started operating? Where is confidence earned through repeated use rather than assumed on the basis of a promising pilot? Alongside these questions sits a related one that the Role Alignment Map is designed to answer: whether the leadership system as a whole shares a common interpretation of where AI is creating value, who owns it, and what the strategic priorities are. Adoption-stage divergence and strategic misalignment are not the same problem, and they do not resolve through the same interventions.
These are the places where budget should follow. Not because activity is highest, but because behaviour has stabilised enough to suggest the value will persist.
The Q1 fieldwork will test whether the concentration pattern anticipated here holds in practice, and whether it is concentrating by role before it concentrates by sector, by company size, or by use case.
That is worth watching.
This essay is part of Quaie’s Founding Essay Series, examining how organisations decide to adopt AI role by role, over time.
Notes and Sources
¹ Marketing and customer service as leading functional areas for stabilised AI use: Consistent with McKinsey Global Survey on AI (2024), which identified customer service, marketing, and software engineering as the three most common functions for generative AI deployment. BCG AI Radar 2025 (January 2025, 1,803 C-level executives) corroborates marketing and customer operations as early value-concentration areas.
² Microsoft Copilot adoption-to-value gap: Microsoft reported 70 per cent of Fortune 500 companies purchasing Copilot licences by late 2024 (Microsoft earnings calls). Gartner found fewer than 5 per cent had moved beyond limited pilot (Gartner research, mid-2024). The gap between platform purchase and stabilised organisational use illustrates why deployment metrics alone do not capture value formation.
Quaie’s six analytical constructs (the Role Shift Index, Role Lead-Lag Ranking, Consensus Formation Time, Role Influence Index, Organisational Adoption Gradient, and Role Alignment Map) are described in full in the forthcoming book The Role Layer: The Missing Intelligence in Enterprise AI Adoption (Quaie Ltd, 2026) and in subsequent essays in this series.



