6 April 2026·5 min readROIAI strategy

How to measure ROI on an AI investment

A practical framework for measuring the return on an AI programme - including the costs most teams forget, and the outcomes most vendors won't commit to.

A
Ajay Dhillon
Founder

AI programmes are easier to fund than to defend. Leadership approves the first budget on a plausible business case. Twelve months later, finance asks what they got for it. The programmes that survive that conversation have measured themselves honestly from day one. The ones that don't, don't.

This is a practical framework for measuring ROI on an AI investment. It is written for the operator or finance partner preparing the case, not for the vendor pitching it.

Start with the outcome, not the output

The most common mistake in AI ROI analysis is measuring outputs rather than outcomes.

  • Output: "We processed 400,000 documents through the AI system."
  • Outcome: "Cycle time fell from 9 days to 4 hours and overtime was eliminated."

Outputs are easy to count. They don't defend a budget. Outcomes are harder to attribute but they are what the CFO needs.

Every AI programme should have a named outcome metric, agreed on day one, with a baseline and a target. If the programme cannot name this metric at scoping, it is not yet ready to be funded.

What goes in the numerator

Outcomes fall into four buckets, roughly in descending order of ease-of- measurement.

1. Cost avoided

The easiest number to defend. Manual hours removed, vendor contracts terminated, overtime not paid. Usually calculated as (hours saved per month × fully-loaded hourly rate × months). Fully-loaded, not headline - include benefits, tools, overhead. The multiplier is usually 1.4–1.8× the salary number.

2. Revenue uplift

Harder to attribute cleanly. Conversion rate improvements, new lines activated, recovered sales from faster response. If you can A/B test, do. If you can't, isolate a market or region as a control.

3. Risk reduced

Often the largest number and the hardest to defend. Fewer compliance incidents, lower fraud rate, reduced reputational exposure. Express as expected annual loss avoided, with a probability assumption you can justify.

4. Capacity created

New capabilities the organisation now has. Not directly monetisable but real. Usually best expressed as "we can now do X, which we couldn't before."

What goes in the denominator

This is where most analyses fall short. The real cost of an AI programme has more line items than a first-draft budget shows.

  • Build cost (obvious).
  • Infrastructure cost (inference, data platform, tooling).
  • Operations cost (on-call, evaluation review, regression work).
  • Governance cost (15–25% of programme cost, often not line-itemed).
  • Retraining and tuning (quarterly at minimum).
  • Integration debt paid by downstream teams (often invisible; real).
  • Opportunity cost of the engineering capacity committed.

A working number for total first-year cost is usually 1.5–2× the original build quote once all of these are included. Budget for that reality.

Unit economics

For production systems at scale, unit economics matter more than top-line ROI. Express the cost per transaction / call / decision / document, and track it monthly.

A healthy trajectory: unit cost falls over time as volume grows, model choices mature, and tuning compounds. An unhealthy trajectory: unit cost stays flat or rises because the system is running on the wrong model tier and the team hasn't optimised.

If you can't quote a unit cost by month six, the programme is not being measured seriously.

How long to wait before measuring

First-year numbers are provisional. Three reasons:

  1. Baselines stabilise at 90 days. The first weeks include launch artefacts and unusual patterns.
  2. Volumes grow. Many AI programmes ship small and ramp. Early ROI numbers reflect a smaller workload than the eventual state.
  3. Governance overhead is front-loaded. The evaluation harness is expensive to build and cheap to run. Year-one costs include both; year- two costs mostly the latter.

A good review cadence is: weekly operational review, monthly outcome review, quarterly ROI review, annual case refresh.

The honest conversation with finance

Three commitments that make the finance conversation easier.

Agree the scoreboard before the first build sprint. A one-page document with the outcome metric, baseline, target, stop condition, and review cadence. Signed by the programme sponsor and the finance partner.

Measure unit cost from day one of production. Not "we'll add cost tracking later." If the system doesn't know what each call costs, the system isn't ready for production.

Write the stop condition in advance. The specific circumstance under which the programme would be rescoped or halted. This is the single strongest signal of a programme being managed rather than hoped for.

Common patterns of fake ROI

Worth naming, because you'll see them in vendor proposals:

  • Full hourly rate × hours automated, with no accounting for the humans still doing the work or the time they now spend on supervision.
  • Revenue attribution via correlation, without controls or holdouts.
  • Cost of the old process overstated, often by treating peak-season staffing as the baseline.
  • Omitted governance cost, which usually arrives later and is pretended not to exist.

A real ROI case survives scrutiny from a finance partner who has not seen the vendor slides. A fake one doesn't.

An example, for reference

One of our deployments automated a claims routing workflow. The case looked like this at year one:

  • Outcome metric: claims cycle time.
  • Baseline: 11 days.
  • Target: 4 days.
  • Actual: 3.2 days (beat).
  • Hours removed: 2,400 per month.
  • Cost avoided: ~$1.6M annualised (fully-loaded).
  • Programme cost, year one: ~$720K including infrastructure and governance.
  • Net year-one ROI: ~1.2×.
  • Unit cost per claim: $0.62.
  • Trajectory: unit cost on track to fall to $0.28 by year two.

Notice what is not in that summary: no "time saved" as output, no uncontrolled revenue attribution, a full denominator including governance, and a stated year-two trajectory. This is what a defended ROI case looks like.

Related reading

Frequently asked

How do you measure ROI on an AI investment? Agree an outcome metric on day one (cost avoided, revenue uplift, risk reduced, or capacity created). Track it weekly. Review monthly. Denominator must include build, run, governance and operations - typically 1.5–2× the original build quote.

How long does it take to see ROI from AI? Most production AI programmes reach break-even between months 9 and 18. Programmes that break even earlier typically under-invested in governance; programmes that never break even typically skipped scoping.

What is a reasonable first-year ROI for an AI programme? Between 1.0× and 2.5× in cost-avoidance terms is typical for a well-scoped programme. Revenue-uplift ROI is more variable and harder to attribute. Anything a vendor promises above 5× in year one without a specific worked example should be treated with skepticism.

What costs do teams forget in AI ROI analysis? Governance, infrastructure, operator time, retraining, and downstream integration work. Together these are usually 40–70% of programme cost and are often missing from the original quote.


Our AI strategy service includes a scoring framework specifically designed to produce defensible ROI cases - not marketing ROI.

Written by
Ajay Dhillon · Founder
10 · Start here

Let’sbuildyoursystemnext.

Thirty minutes with someone who’d be doing the work. No slide deck, no intake form. We’ll tell you what’s feasible, where you’ll hit friction, and what we’d pick up first.

Response
< 24 hours
First read
No NDA needed
Bangalore / Remote
UTC ±12