AskAjay.ai
AI Strategy14 min read · February 8, 2024

Measuring AI ROI: Why Most Count the Wrong Things

Three measurement frameworks — enhanced cost-benefit, balanced scorecard, and predictive modeling — for capturing AI value that traditional ROI calculations miss. Includes a four-step playbook for boards.

Most AI ROI calculations miss half the value. Three measurement frameworks and a four-step playbook for executives who need numbers boards trust.

Ajay Pundhir
Ajay PundhirAI Strategist & Speaker
Share
AI Strategy

Measuring AI ROI: Why Most Count the Wrong Things

Key Takeaways

  • 42% of businesses cannot even define what ROI means for their AI projects
  • The "Often Missed" value category typically accounts for 40-60% of total AI value
  • Killing AI projects at 6 months destroys initiatives that would succeed at 18 months
  • AI-savvy boards outperform peers by 10.9 percentage points in return on equity

Last year, a $400 million industrial company asked me to evaluate their AI portfolio. The CFO had one number: $3.2 million spent. The CTO had model accuracy percentages. The head of operations pointed to a pilot that “seemed to be working.” The board wanted a simple answer: is our AI investment paying off?

Three weeks of digging turned up a split result nobody expected.

Their two flagship projects — the ones getting executive airtime and internal champions — had negative ROI by any traditional measure. Meanwhile, a quiet document-processing workflow buried in operations was saving $1.7 million annually. Nobody counted it because it didn’t fit the “transformative AI” narrative the board preferred to hear. The unsexy project was the winner. The showcase projects were drains.

They’re not alone. A 2023 PwC survey found that 42% of businesses can’t even define what ROI means for their AI projects. Not struggling to achieve it. Struggling to define it.

The question is no longer ‘Can AI do this?’ It’s ‘How well, at what cost, and for whom?’

— Stanford Human-Centered Artificial Intelligence, AI Index Report 2024

The executives who get AI ROI right aren’t the ones with better spreadsheets. They’re the ones who’ve accepted that measuring AI requires a vocabulary the finance team doesn’t have yet — and they’re willing to build it.

Why Traditional AI ROI Frameworks Fail

Traditional ROI is beautifully simple: (Gain − Cost) / Cost. Works perfectly for buying a machine on a factory floor. Falls apart completely for AI.

I learned the reasons the hard way, through three advisory engagements that taught me more from getting the measurement wrong than from any framework I’d read.

The Isolation Problem

AI rarely operates in a vacuum. Deploy an AI-powered recommendation engine alongside a website redesign, and a 15% sales increase follows. Was it the AI? The UX improvements? A seasonal trend? A/B testing helps, but real-world deployments are far messier than controlled experiments.

One retail client spent six months arguing about whether their AI personalization engine or their redesigned checkout flow drove a 22% conversion increase. Six months. By the time they settled the attribution question, their competitor had shipped both and moved on to the next thing. That engagement taught me something I now repeat to every client: perfect attribution is the enemy of informed action. You don’t need to know which raindrop filled the bucket. You need to know the bucket is filling.

The Intangibility Trap

A 2024 Deloitte survey found that 74% of companies see customer service and experience as their top area for AI returns. Try putting “better customer experience” on a P&L statement.

Improved decision-making. Reduced cognitive load on your analysts. Faster time-to-insight for your product teams. These are real outcomes that compound over quarters. In regulated industries like healthcare, AI’s risk-reduction value can dwarf direct cost savings — but that value never appears in a traditional ROI calculation. (If your AI touches patient data, the HIPAA compliance considerations become non-negotiable regardless of how you measure returns.)

The Time Horizon Mismatch

Most AI projects follow a J-curve. Significant upfront investment, a painful stretch of learning and optimization, then accelerating returns. But executive planning cycles operate on 12-month horizons. You’re judging a marathon runner at the 100-meter mark.

That industrial client I mentioned? Their document-processing AI took fourteen months to reach positive ROI. Fourteen. If they’d reviewed it at six months — which was the original plan — they would have killed a project now saving $1.7 million a year. I think about that near-miss often.

The AI Investment J-Curve

Why timing your measurement matters as much as the measurement itself

1
Cost
Months 0–6: Investment Phase

Infrastructure, data preparation, talent acquisition, model development. Costs are high and visible. Returns are near zero.

2
Development
Months 6–12: Learning Phase

Model training, integration testing, user adoption struggles. Early gains appear but fall short of justifying the spend. This is where most kill decisions happen — and most of them are premature.

3
Growth
Months 12–18: Optimization Phase

Model refinement, tighter workflow integration, compounding improvements. ROI begins to materialize. The team finally stops defending the project and starts scaling it.

4
Returns
Months 18+: Value Realization

Scaled deployment, network effects, organizational learning compounds. True ROI emerges and frequently exceeds initial projections by 2–3x.

If your AI ROI review happens at six months, you’re almost guaranteed to kill a project that would have succeeded at eighteen. I’ve watched it happen three times in the last two years. Each time, the post-mortem revealed the same thing: premature measurement, not a bad investment.

Three Frameworks That Actually Work

None of this means AI ROI is unmeasurable. It means you need different instruments.

After getting the measurement question wrong more times than I’d care to admit, I’ve landed on three approaches. They solve different problems. You’ll probably need all three, which is inconvenient but unavoidable.

Enhanced Cost-Benefit Analysis

Start with the traditional approach but blow out the scope. The mistake I see in nearly every engagement: teams stop at the obvious costs and benefits. In my experience, the “Often Missed” column below typically accounts for 40–60% of total value.

Let that number land for a second. Half the value of your AI investment might be invisible to the spreadsheet your CFO is using to evaluate it.

AI Cost-Benefit Analysis: The Full Picture

Most companies only track the left column

Cost CategoryOften MeasuredOften Missed
Infrastructure

Hardware, cloud compute, software licenses

Opportunity cost of delayed deployment; technical debt from legacy integration
Data

Acquisition, cleaning, storage

Ongoing data quality maintenance; privacy compliance costs; data pipeline reliability
Talent

Hiring data scientists, ML engineers

Upskilling existing workforce; change management; productivity dip during transition
Value

Direct cost savings, revenue increases

Employee satisfaction; decision quality improvement; competitive positioning; risk reduction

Capgemini’s 2024 generative AI research documented this across their enterprise deployments. One AI-powered invoice processing case tracked results for twelve months: 70% reduction in processing time, 30% reduction in direct costs. Solid numbers. But the full value story — error rate improvements, employee reallocation to higher-value work, vendor satisfaction gains — added another 40% that traditional ROI would have missed entirely.

The California Management Review formalized this into three ROI categories: traditional financial ROI, intangible ROI (trust, reputation, employee morale), and real option ROI — the strategic flexibility AI creates for future investments. Most organizations only measure the first category. The Governance Playbook covers how to operationalize all three.

The Balanced Scorecard for AI

When I first started recommending balanced scorecards for AI measurement, most CIOs pushed back. “That’s a 90s framework,” they’d say. Kaplan and Norton, 1992. Fair point about the vintage.

Completely wrong about the relevance.

The balanced scorecard is tailor-made for AI measurement precisely because it forces you beyond financial metrics. It makes you track what matters even when the CFO’s spreadsheet can’t capture it.

Bank of America’s Erica chatbot is the case I point to most. They don’t just measure cost savings from reduced call center volume. They track customer satisfaction scores for chatbot interactions against phone interactions. Per their October 2024 press release: over 2 billion client interactions since launch, with 98% resolution rates. That’s not just efficiency. It’s a measurably better experience for the customer — the kind of value that evaporates if you’re only looking at cost lines.

The balanced scorecard forces a question most AI teams avoid: what does success look like beyond the spreadsheet? If your measurement framework can’t answer that, swap the framework.

Predictive AI ROI Modeling

Given the J-curve reality, predictive modeling lets you forecast ROI before you’ve committed the full budget. I use this approach with every advisory client evaluating an AI investment above $500K. Below that threshold, the modeling overhead usually isn’t worth it.

Four Components of Predictive AI ROI Modeling

Structured estimation, not guesswork

Data

Start with what’s already happened in your sector. If industry peers have deployed similar AI, their 6/12/24-month data is your best starting point. No internal benchmarks? McKinsey’s State of AI reports and Deloitte publish sector-level data that’s good enough to sanity-check your projections. Don’t build your model on hope.

Financial

This is where teams consistently underestimate. Detailed infrastructure, data, and talent cost projections must account for the J-curve trajectory. Budget for the costs everyone forgets: data quality maintenance, model retraining, compliance overhead, and the organizational drag of change management. In engagements I’ve led, these “hidden” costs add 25–40% to initial projections. Every single time.

Risk

Ask the uncomfortable questions. How much does projected ROI shift if adoption lands 20% lower than expected? What if data quality delays the timeline by three months? What if your best ML engineer leaves in month four? Most ROI projections assume everything goes to plan. Yours shouldn’t. HBR’s research on AI project management confirms that the organizations that model downside scenarios are far more likely to deliver on AI investments than the ones that only model the upside.

Strategy

Give leadership a range, not a false-precision point estimate. Boards respond far better to “$1.2M–$2.8M over 24 months depending on adoption” than to “$2.1M ROI.” The honest range builds more trust than the confident number. I have never once had a board lose confidence because I gave them a range. I have seen several lose confidence when a point estimate turned out to be wrong.

The shift predictive modeling creates is subtle but important. It moves the executive conversation from “Will this work?” to “Under what conditions does this work, and how do we create those conditions?” That second question is dramatically more productive. It connects directly to the AI Use Case Canvas, which structures exactly those conditions.

A Playbook for Boards That Want Real Numbers

These three frameworks aren’t mutually exclusive. They’re cumulative. Start with enhanced cost-benefit analysis for your financial foundation. Layer on balanced scorecard metrics for the value the spreadsheet misses. Before committing real budget, run predictive models to stress-test assumptions.

The sequence matters. Skip the order and you’ll have metrics without context.

Four Steps to AI ROI That Boards Actually Trust

From measurement chaos to executive confidence

Step 1

Identify 3–5 KPIs that map to specific business outcomes. At minimum: one financial metric, one operational metric, one stakeholder metric. If you can’t define success before you start, you can’t measure it after. The AI Use Case Canvas provides a structured approach — I designed Block 4 (Value Proposition) and Block 10 (Unit Economics) specifically for this purpose.

Step 2

Data pipelines for your AI models are useless without data pipelines for your metrics. Build dashboards that track KPIs from day one — not as an afterthought at review time. Budget 10–15% of your total AI investment for measurement and monitoring. That sounds high. It’s not. The RAG evaluation framework shows how this plays out in practice for retrieval-augmented generation deployments.

Step 3

You don’t need executives who can code. You need executives who ask the right questions. What does this metric actually tell us? What’s the counterfactual — what happens if we do nothing? Are we measuring activity or impact? MIT Sloan research shows AI-savvy boards outperform peers by 10.9 percentage points in return on equity. That’s not a rounding error. The 5-Pillar AI Readiness Assessment includes a leadership capability dimension for this reason.

Step 4

Stop making single go/no-go decisions on AI investments. Structure them as stages with measurement checkpoints: Pilot → Validate → Scale. Each stage has defined metrics that must be met before releasing the next tranche of budget. This de-risks the investment and — just as important — builds organizational confidence that AI can be governed. For founders operating with less organizational maturity, the Responsible AI Playbook adapts this approach for earlier-stage companies.

The biggest ROI mistake I see in the field: treating measurement as a one-time event. A quarterly board deck isn’t measurement. The companies that win at AI measure weekly, iterate monthly, and report quarterly. Continuous, not episodic.

Where This Breaks Down

Any honest framework comes with caveats. Mine is no exception.

Early-stage startups will find these frameworks assume organizational maturity they don’t have yet. If you’re pre-revenue, you need a leaner approach focused on product-market fit — balanced scorecards are overkill. The Incubators & Accelerators Guide covers approaches better suited to that stage.

Breakthrough AI research doesn’t fit either. When the goal is fundamental capability development rather than deployment, ROI framing itself is the wrong lens. Basic research doesn’t follow a J-curve. It follows a step function — nothing, nothing, nothing, breakthrough.

Speed-of-disruption scenarios present a different problem entirely. Sometimes the competitive cost of not investing in AI dwarfs the ROI calculation. If your industry is being disrupted by AI-native competitors, waiting for perfect measurement is its own form of risk — potentially the largest one on the table. As the 2026 AI forecast notes, CFOs are increasingly taking over AI strategy from CTOs precisely because measurement gaps have become board-level concerns.

And then there’s measurement theater — the most dangerous failure mode. Teams game the metrics, produce impressive dashboards, and the numbers look great while the underlying deployment stalls or quietly erodes trust with users. I’d rather you acknowledge these blind spots upfront than discover them at a board meeting.

ROI depends on readiness. Assess your readiness across 5 pillars — organisations with balanced readiness see 3x higher AI ROI.

What This Means for Your Next Board Meeting

AI ROI isn’t unmeasurable. It’s differently measurable. The executives who succeed are the ones who accept three uncomfortable truths:

  1. Traditional ROI frameworks miss intangible benefits that compound over time. In my advisory engagements, the “Often Missed” category in the cost-benefit matrix typically represents 40–60% of total realized value. Invisible to anyone using only financial metrics.
  2. Measurement timing matters as much as methodology. Evaluate too early and you kill projects that would have succeeded. Build the J-curve into your expectations from day one — and protect early-stage projects from premature review cycles.
  3. The best metric is the one your entire leadership team agrees on before the project starts. Not the one that looks best in the post-mortem.

The $13 trillion AI opportunity McKinsey projected isn’t going to materialize through better models alone. It’s going to materialize through better measurement — the kind that connects AI capabilities to business outcomes with rigor, patience, and intellectual honesty. IBM’s research reinforces this: organizations investing in AI ethics and governance alongside deployment achieve 34% higher operating profit margins from their AI initiatives. Governance isn’t overhead. It’s a multiplier.

Stop asking “What’s the ROI of AI?” Start asking “What’s the ROI of this specific application, measured against these specific outcomes, over this specific timeframe?” The precision of the question determines the usefulness of the answer.

Frameworks to Put This Into Practice

The AI Use Case Canvas builds ROI thinking into every stage of AI evaluation — from value proposition through unit economics. For regulatory contexts where compliance costs reshape your ROI equation, the GDPR compliance guide and Minimum Viable Governance framework cover the governance dimensions boards increasingly expect to see quantified.

To assess whether your organization has the foundations to measure and realize AI value, the 5-Pillar AI Readiness Assessment is the diagnostic I use with every new advisory client. For the broader investment landscape these measurements exist within, the 2026 AI Forecast covers the macro trends shaping every ROI calculation this year.

If you’d like a facilitated ROI assessment for your leadership team — the kind that surfaces the “Often Missed” value and builds a measurement framework your board will actually trust — that’s what the advisory practice is for. I also cover measurement frameworks in depth in the AI Strategy for Leaders curriculum, for teams that want to build the capability internally.


Ajay Pundhir
Ajay Pundhir

Senior AI strategist helping leaders make AI real across four continents. Forbes Technology Council member, IEEE Senior Member.

Let's Talk

Get Weekly Thinking

Join 2,500+ leaders who start their week with original AI insights.