GALAHAD
/Articles
All ArticlesHomeContactWork With Us →
Galahad/Articles/AI Strategy
AI Strategy

The 90-Day AI Audit: What to Measure, What to Ignore

Last updated 2026-03-20

After 90 days, most AI projects have either proven their value or revealed their flaws. But most organisations measure the wrong things. They track usage instead of outcomes, count logins instead of decisions improved, and report adoption rates that tell you nothing about whether the AI is actually earning its place. Here's how to separate signal from sunk cost.

Why 90 days matters

Ninety days is long enough to get past the novelty effect and short enough to course-correct before you've committed serious resources. It's the window where an AI pilot either demonstrates genuine value or quietly becomes shelfware that nobody admits isn't working.

The problem is that most organisations don't design their pilots with a 90-day evaluation in mind. They launch, they monitor loosely, and after three months they have usage data but no evidence of impact. The decision about whether to continue, expand, or kill the project becomes political rather than analytical.

A well-designed 90-day audit starts before the pilot begins. You define the success criteria upfront, instrument the system to measure them, and commit to an honest evaluation at the end - including the possibility that the answer is 'stop.'

Metrics that actually predict success

The metrics that predict long-term AI success are not the ones most organisations track. Usage and adoption tell you whether people are logging in. They don't tell you whether the AI is making a difference.

The metrics that matter are decision quality (are better decisions being made?), time-to-decision (are decisions happening faster?), error reduction (are fewer mistakes being made?), and capability transfer (is the team learning from the AI, not just depending on it?).

Each of these requires a baseline. If you didn't measure decision quality before the AI was deployed, you can't measure the improvement. This is why the audit design has to happen before the pilot launches - not three months later when someone asks 'so, is it working?'

Vanity metrics that fool boards

Adoption rate is the most dangerous vanity metric in AI. A high adoption rate means people are using the tool. It doesn't mean the tool is delivering value. An AI assistant with 90% adoption and zero measurable impact on business outcomes is an expensive habit, not a strategic asset.

Other vanity metrics to watch for: number of queries processed (activity, not impact), user satisfaction scores (people like novelty), time saved per task (often self-reported and inflated), and 'AI-influenced revenue' (a category so broad it's meaningless).

None of these are worthless. But none of them answer the only question that matters: is this AI system making the organisation measurably better at something that matters to the business?

How to isolate the contribution of AI

The hardest part of any AI audit is attribution. When outcomes improve after deploying AI, how do you know the AI caused the improvement? Maybe the team got better. Maybe the market shifted. Maybe the process changes that accompanied the AI deployment were the real driver.

Perfect attribution is usually impossible. But good-enough attribution is achievable with simple controls: A/B comparisons where feasible, before-and-after measurement against stable baselines, and qualitative assessment from the people closest to the work.

The goal isn't scientific certainty. It's enough evidence to make a confident decision about whether to invest further, adjust course, or stop. If after 90 days you can't articulate how the AI has improved a specific outcome with reasonable evidence, you have a problem - and more time probably won't fix it.

From pilot to production

The 90-day audit isn't just an evaluation - it's the decision point for what comes next. There are three possible outcomes, and each requires a different response.

If the pilot has demonstrated clear value against the metrics you defined, the question is how to scale. Scaling is its own challenge - what works for one team doesn't automatically work for ten - but you have a foundation of evidence to build on.

If the pilot has shown partial value but hasn't met its success criteria, the question is whether to adjust or stop. Sometimes a pivot - different use case, different team, different scope - unlocks the value. Sometimes the honest answer is that the AI isn't the right tool for this problem.

If the pilot has failed, the question is what to learn. Failure isn't wasted investment if it produces genuine insight about what doesn't work and why. The worst outcome isn't a failed pilot - it's a failed pilot that gets extended because nobody wants to admit it didn't work.

Frequently Asked Questions
When should I start planning the 90-day audit?
Before the AI pilot launches. The audit design - success criteria, baseline measurements, instrumentation - should be part of the pilot plan from day one. If you wait until the pilot is running to decide what to measure, you've already lost the ability to evaluate it properly.
What if our AI pilot shows high adoption but low measurable impact?
High adoption with low impact is a warning sign, not a success metric. It usually means the AI is being used for low-value tasks or that the impact metrics aren't aligned with business outcomes. Investigate what people are actually using it for and whether that aligns with the original business case.
How do I present a failed AI pilot to the board?
Frame it as a learning investment, not a failure. Present what you tested, what you measured, what didn't work and why, and what you'd do differently. A well-documented failed pilot with clear learnings is more valuable to a board than a vague success story with inflated metrics.
Can Galahad help with AI audits?
Yes. The Galahad Diagnostic includes an AI maturity assessment and can be structured as a 90-day audit framework. We help define success criteria, design measurement systems, and provide honest evaluation at the end - including recommendations to stop if that's the right call.
Related Articles

The Board AI Briefing: What to Say, What to Leave Out

Every CFO and board member has the same three questions about AI. Get the framing wrong and you'll spend the next year defending ROI. Get it right and you move fast.

Read article →

Over 1,000 Senior Leaders Trained. Here's the Thing They All Struggled With.

Running AI enablement training across 1,000+ senior leaders clarifies something fast: the hard part isn't capability. It's deciding where AI stops and human judgement begins.

Read article →

Stop Building AI You Should Be Buying

90% of organisations building custom AI are wasting money on commodity problems. The build-vs-buy decision isn't a cost calculation — it's a competitive advantage test that most fail.

Read article →

Want to go deeper?

If this article raised questions about your own AI strategy, we're happy to talk it through. No pitch. No pressure.

Start a Conversation →

This article provides general information and opinion. It does not constitute legal, financial, or technical advice. Always consult qualified professionals for decisions specific to your organisation.

Galahad
AI that knows its place. · Founded by Ross Barnes
hello@galahadgroup.co.uk
HomeServicesArticlesikigAIEnableGrailContact
© 2026 Galahad. All rights reserved.London · Global