You don't have a tooling problem. You have a single-source-of-truth problem

A founder and operations lead at a desk with three laptops, printed reports, and mugs, leaning in studying one screen together
TL;DR

The visible symptom of tool sprawl is the SaaS bill. The structural problem is fragmented ownership of the data, which makes every important number contestable in the senior team. Consolidating vendors doesn't fix it. Building a single source of truth does, and the work has to happen before any AI or analytics ambition lands, or those investments stall on inputs nobody trusts.

Key takeaways

- The SaaS bill is the surface signal. The real cost of tool sprawl is the time, decisions, and senior-team trust lost when no two systems agree on the same number, which more than half of SMB leaders report living with weekly. - The numbers underneath are bigger than founders imagine. Workers spend 102 minutes a day searching for information and lose five working weeks a year to context switching, on top of the proportion of agency staff who say their tech stack actively hindered them. - The trust gap is inverted at the top. Forty-three percent of C-level executives find their information unreliable, against thirty-two percent of more junior staff. The people making the largest calls have the least confidence in the inputs. - Cutting from twenty-three subscriptions to twelve still leaves eight versions of utilisation. Tool sprawl is the symptom. The disease is fragmented ownership of the metrics that drive decisions. - AI on fragmented data is the most expensive way to discover the data was the problem. Single source of truth comes first; AI and analytics plug in afterwards, or they don't get used.

Sixty people, twenty-three SaaS subscriptions, and an ops lead who builds the weekly report by hand every Monday morning by pulling from six places. The numbers in that report disagree with the numbers in the CRM, which disagree with the numbers in the finance pack the accountant sends over. Every Monday standup spends fifteen minutes arguing about which version of utilisation is real. Three smart people, three plausible answers, no resolution. Nobody is wrong. Everybody is reading from a different source.

The founder feels this as tool sprawl. The bill arrives, the bill keeps growing, and the instinct is to consolidate vendors. That’s the wrong target. The SaaS bill is small theatre next to what the firm actually loses every week to data nobody fully trusts.

Why is your SaaS bill the wrong number to be looking at?

The bill is the visible cost of tool sprawl. The actual cost is what your team can’t decide because no two systems agree on the same number. More than half of SMB leaders report frequent data inconsistencies caused by silos, which makes the data argument the median experience, not a fringe complaint. The bill, against the time and decisions lost to that argument, is a rounding error.

The numbers underneath this are larger than founders typically imagine. The average employee spends 102 minutes a day searching for information needed to do the job. Five working weeks a year per knowledge worker get lost to context switching, with about 40 per cent of productive time consumed by chronic multitasking. In a forty-person firm, that’s the equivalent of two full-time staff who exist to chase information that should already be agreed.

Where the trust deteriorates first is at the top. Forty-three percent of C-level executives find their information unreliable, against thirty-two percent of more junior staff. That’s the worst possible inversion. The people making the largest decisions have the least confidence in the inputs. Across an agency benchmark, 33 per cent of staff said their tech stack had no productivity impact and 14 per cent said it actively hindered them, which is a remarkable indictment of a category many firms now spend more on every year.

The cumulative cost lands as a hesitation tax. Decisions get deferred because the data feels soft. Forecasts get hedged. Hiring slows because nobody can confirm the utilisation picture. Pricing reviews get pushed because the margins don’t reconcile cleanly. None of that shows up on the SaaS invoice. All of it is what happens when the senior team can’t trust their own data.

Why doesn’t consolidating tools fix it?

You can cut from twenty-three subscriptions to twelve and still have eight versions of utilisation. Consolidation is the move founders reach for first, and it’s the move that disappoints most reliably. The subscription fee shifts; the underlying confusion stays. The reason is structural. Tool sprawl is the surface symptom. The disease is fragmented ownership of the data that drives decisions, which doesn’t get fixed by changing the labels on the systems.

The pattern that creates sprawl in the first place explains why consolidation rarely undoes it. Tools get added under pressure with no central oversight. Sales needed a CRM, so Sales bought one. Finance needed a different general ledger, so Finance bought one. Operations spun up a project tracker. Marketing added an email platform. Each is a sensible local decision. None is owned at the firm level. Twenty-three sensible local decisions produce one collective mess, and merging some of them into a “unified” platform doesn’t change who owns the underlying definitions.

A useful test is to ask, for any contested metric, who decides what it means and where the canonical number comes from. In the typical owner-led firm the honest answer is “nobody, or the founder, depending on who’s in the room”. That answer is the actual problem. Until somebody owns the question, new tools just create new silos at lower cost.

What does a single source of truth actually mean here?

A single source of truth is a small, deliberate set of agreed definitions, named data sources, and one canonical pipeline for the metrics that drive decisions. It is not a tool and it is not a dashboard, even though both might be involved. The high-performing version is what some firms call a metrics handbook: each metric defined once, sourced from one system, reported on one cadence, and tied to the decisions it actually informs.

Scoro put the underlying point well: growth doesn’t fall apart because you don’t have enough leads, it falls apart when no one agrees on what the business data actually means.

Take “utilisation” as the worked example. Definition: billable hours divided by available hours, both measured at week granularity. Canonical source: the time-tracking system, with the explicit rule that hours not in time tracking don’t exist for this calculation. Cadence: weekly, in a single place, owned by one person. Decision relevance: under sixty-five percent triggers a hiring pause; over eighty-five percent triggers a hiring conversation. That’s one metric pinned down. Repeat for project margin, pipeline value, cash position, retention, and capacity, and the senior team has a spine the rest of the data can hang off.

The crucial point is that the handbook is the thing that doesn’t move. Tools come and go, but the canonical definition stays put. Once that’s true, tool consolidation becomes a question of fit rather than a panicked response to feeling overwhelmed, and many firms find they can rationalise their stack calmly because the metrics they care about have a fixed home.

Why does this have to happen before any AI work?

AI on top of fragmented data is the most expensive way to discover the data was the problem. Founders who bolt analytics or AI onto the existing mess end up one of two ways. They spend heavily to clean the inputs as part of the project, often more than the AI itself costs. Or they ship outputs nobody acts on, because the numbers contradict the version the senior team carry in their heads.

The McKinsey state-of-AI numbers carry this neatly. 88 per cent of organisations now use AI in at least one function. Only 39 per cent report enterprise-level profit impact. The Boston Consulting Group analysis of “future-built” firms shows the same shape from a different angle: the five per cent of firms generating outsize value share a clean, owned data foundation, while the rest run pilots that don’t compound. The differentiator is the data foundation, not the model.

The order, then, is single source of truth first, AI second. With agreed definitions in place, a churn model that uses retention numbers gets used. A pricing recommendation built on the canonical project margin pipeline gets adopted. A capacity forecast drawn from the official utilisation source gets included in the hiring conversation. Without that work in place, the AI works as advertised, nobody acts on the outputs, and the investment delivers a technical success and a commercial nothing.

The same pattern applies to ordinary analytics. Dashboards built on canonical sources get looked at. Dashboards drawing from one of three competing sources get ignored, because the senior team has been trained that data they can’t defend in the room is data they don’t act on.

What this gives you back

A senior team that trusts its own numbers behaves differently. The Monday standup stops being a fifteen-minute argument and starts being a twenty-minute decision. Hiring conversations resolve. Pricing reviews land. AI projects, when you do them, plug into something solid and produce outputs the firm can act on rather than reports the firm can ignore.

The handbook itself is short. Many firms write a v1 in two focused days. The discipline is in maintaining it as the business changes, which means assigning it an owner. Without one, the silos return, the standup goes back to arguing about utilisation, and the next round of tool consolidation starts to look attractive again for the wrong reasons.

If the standup pattern feels familiar, book a conversation. The diagnostic work usually takes a morning, and the metrics handbook v1 is rarely as far away as founders fear.

Sources

- McKinsey & Company (2025). The State of AI: Global Survey. 88 per cent of organisations now use AI in at least one function, but only 39 per cent report enterprise-level EBIT impact, the gap that fragmented data tends to sit inside. https://www.mckinsey.com/capabilities/quantumblack/our-insights/the-state-of-ai - Boston Consulting Group (2025). Are You Generating Value from AI? The Widening Gap. The five per cent of "future-built" firms achieving 5x revenue gains and 3x cost reductions of peers; data foundation cited as a recurring differentiator. https://www.bcg.com/publications/2025/are-you-generating-value-from-ai-the-widening-gap - MIT Sloan Management Review (2024). The Cultural Benefits of Artificial Intelligence in the Enterprise. Why AI deployments stall when underlying data ownership and definitions are not in place across the senior team. https://sloanreview.mit.edu/projects/the-cultural-benefits-of-artificial-intelligence-in-the-enterprise/ - Harvard Business Review (2022). Why Data Culture Matters. Empirical work on the link between agreed data definitions, decision velocity, and the gap between firms that get value from analytics and those that don't. https://hbr.org/2022/05/why-data-culture-matters - MIT Center for Information Systems Research (2024). Data foundations and AI value. Research on why AI investment fails to convert when underlying data definitions and ownership are not in place. https://cisr.mit.edu/publication/2024_0801_AIDataFoundations - Starmind (2024). The Future of Work report. 102 minutes a day spent searching for information, $71m annual productivity loss in large organisations, 43 per cent of C-level vs 32 per cent of junior staff find information unreliable. https://www.starmind.ai - Leroy, S. (2009). Why is it so hard to do my work? The challenge of attention residue when switching between work tasks. Organizational Behavior and Human Decision Processes, 109(2), 168-181. The foundational research on attention residue and the cost of switching between tasks, underpinning the five-working-weeks-a-year context-switching estimate. https://www.sciencedirect.com/science/article/abs/pii/S0749597809000399 - NinjaOne (2024). From tool sprawl to a unified tech stack. The tool evaluation trap, departmental tool buying, and the absence of central oversight as the structural drivers of sprawl. https://www.ninjaone.com/blog/from-tool-sprawl-to-unified-tech-stack - PCI (2024). Agency operations: 7 challenges hurting profitability. 33 per cent of agency staff say their tech stack had no productivity impact and 14 per cent say it actively hindered them. https://pci.us/agency-operations-7-challenges-hurting-profitability - Scoro (2024). Misaligned metrics: why growth doesn't fall apart from a lack of leads. The metrics handbook framing and "growth falls apart when no one agrees on what your business data actually means". https://www.scoro.com/blog/misaligned-metrics

Frequently asked questions

I keep being told to consolidate my SaaS stack. Why isn't that the answer?

Because the issue isn't the number of tools. It's that nobody owns the question of what each metric means and where its canonical version lives. Cutting from twenty-three subscriptions to twelve usually leaves the same eight versions of utilisation, just spread across fewer logos. Consolidation can be sensible later, but it's the wrong first move and rarely solves what's actually frustrating you.

Where do I start if I want a single source of truth without rebuilding the business?

With the six to ten metrics that actually drive decisions, not all of them. Pick the ones the senior team argues about most often: utilisation, project margin, cash, pipeline value, retention. For each, agree the definition, name the canonical system, and pick one reporting cadence. Two focused days gives you a v1. The discipline is assigning an owner who keeps it accurate as the business shifts.

How does this connect to AI? I keep hearing both at once.

They have to happen in order: single source of truth first, AI second. AI on fragmented data either pauses while you spend heavily to clean the inputs, or ships outputs that the senior team won't trust because the underlying numbers contradict each other. With agreed definitions and canonical sources in place, the AI plugs into something stable and the outputs actually get used.

This post is general information and education only, not legal, regulatory, financial, or other professional advice. Regulations evolve, fee benchmarks shift, and every situation is different, so please take qualified professional advice before acting on anything you read here. See the Terms of Use for the full position.

Ready to talk it through?

Book a free 30 minute conversation. No pitch, no pressure, just a useful chat about where AI fits in your business.

Book a conversation

Related reading

If any of this sounds familiar, let's talk.

The next step is a conversation. No pitch, no pressure. Just an honest discussion about where you are and whether I can help.

Book a conversation