She has agreed the work is needed. She has read the foundation posts, recognised three of the four exposures in her own firm, and accepted that the next AI tool will fail at the data layer if nothing changes. Now she has thirty minutes and one question. What does the plan look like.
The honest version is short. Ninety days, three phases of about a month each, no new hires and no specialist software at the outset, a maintenance pattern at the end that costs the team about two hours a week. The work is real and it requires discipline, but it fits on a single page and it ends on a known date. Open-ended data programmes lose to revenue work every quarter. A bounded sprint with a firm deadline at day ninety is the version that actually finishes.
Why does a ninety-day frame work better than an open-ended programme?
Indefinite improvement work has no completion date, no scope discipline, and nothing to prioritise against. Gartner finds 57 per cent of companies judge their own data unfit for AI, yet remediation investment stays low because remediation has no end. MIT Project NANDA reports 95 per cent of generative AI deployments produce zero measurable return, with the failures sitting in data and workflow rather than the model. The fix is a firm deadline.
Ninety days is long enough to deliver substantive work and short enough to keep urgency. Owner-operated firms already run monthly accounts and a quarterly review of some kind. Anchoring the sprint to that rhythm means the work integrates with operating cadence rather than competing with delivery. Johanna Rothman’s writing on timeboxing makes the point well. Once the box is fixed, the question stops being when, and becomes what can we realistically do in the time we have.
What happens in phase one, the audit, days 1 to 30?
A one-week structured inventory of where data and knowledge live, followed by a prioritisation workshop, followed by a one-page scope charter that names the three to five issues to be fixed in the next sixty days. The output at day thirty is not detail, it is clarity. The owner can walk a team meeting through what is broken, why each item matters, what fixing it will require, and who is responsible for each workstream.
Week one is the rapid assessment. One or two people, usually the operations lead and a technical generalist, spend fifteen to twenty hours answering four questions. Where does business-critical data live, which systems and which spreadsheets. Who owns each source. Where are the visible inconsistencies. Where is information most often sought and hardest to find. The Future Processing framework gives a useful six-dimension structure for the assessment, but the deliverable is small, a one-page inventory that gets walked through with the owner in a ninety-minute session.
Week two is the prioritisation workshop. Three to four hours. Each issue scored on frequency, business impact and tractability inside sixty days. The arithmetic is simple, frequency times impact divided by a tractability factor, and it surfaces the three to five items that are both critical and achievable. The charter that comes out names each item, defines what resolved looks like by day sixty, and assigns an owner. Weeks three and four turn that into one-page workstream plans, with expected hours and the resources required.
What gets built in phase two, the foundation, days 31 to 60?
The four workstreams the foundation phase typically produces are customer record deduplication, a working business glossary, video walkthroughs of high-risk processes, and a naming convention rolled out from a chosen start date. Not every sprint runs all four. The scope charter from phase one decides which three to five take priority. The pattern below is what owner-operated services firms commonly land on.
Customer deduplication is the first visible win. A typical small business holds three to seven versions of each major entity across CRM, billing, project tooling and accounts. Profisee’s work on duplicated master data documents the revenue and margin friction that follows. Within the sixty-day window the goal is not perfect deduplication, that is a continuing task. The goal is removal of the obvious duplicates, a documented source-of-truth record, and a monthly monitoring report to catch new duplicates at source.
The glossary is concurrent. Five to ten critical entities, customer, product or service line, revenue transaction, staff resource, project, with a business definition for each, the source system, the technical field, the usage rule and the named steward. Atlan’s research on data glossaries makes the design point clearly. A glossary pays back only when it is embedded in the reports and dashboards the team already uses, not when it sits as a standalone document nobody references.
The SOP readiness review is a focused assessment of the top five to seven processes, sorted by risk and documentation state. Category one processes, high risk and known to one or two people, get a ten-minute video walkthrough from the incumbent, transcribed into a checklist with screenshots, then piloted by a different team member while the incumbent observes silently. The naming convention is the lightest workstream. A one-page document, a thirty-minute team training, a chosen start date, and a spot-check over the following month. The ICO’s information governance guide for small business gives the proportionate baseline. Harvard’s biomedical data management guidance reinforces the rule that conventions are set before content is created, not retrospectively.
What does phase three, the discipline, days 61 to 90, actually deliver?
A named owner for each artefact, a quarterly review built into the existing business rhythm, and a five-minute data health item in the weekly operations huddle. The transition the phase is trying to make is from project to operating rhythm. Data governance change management research consistently shows that this is where initiatives die, policies created with enthusiasm and then gradually ignored because nobody has explicit responsibility for them.
Named ownership is the first step. For each artefact, the customer master, the glossary, the documented processes and the naming convention, one person is the steward. The responsibilities are narrow and bounded, respond to questions, maintain version history, communicate changes, spot-check accuracy monthly or quarterly. In a 5 to 50 person business this often collapses to the operations or finance manager, sometimes with help from a technical lead. The deliverable is a one-page stewardship charter that names the owners and the cadence, signed off by the owner-operator and shared with the team.
The quarterly review locks ninety minutes into the calendar in the first or second week after each quarter end. The agenda is short. Duplicates found and remediated, glossary terms questioned or updated, one or two documented processes sampled by someone who did not write the documentation, files spot-checked against the naming convention. The tone is pragmatic. Eighty per cent adoption of the naming convention is good progress. Three glossary terms needing clarification is normal. The weekly operations huddle carries the third mechanism, a rotating five-minute data health item that keeps the work visible without becoming a separate programme. By day ninety the sprint has ended and the operating rhythm has started.
What does good look like at day 90, and what is the next move?
Good looks like data and knowledge that does not block the next AI use case the business takes on. Customer records deduplicated with documented rules. A glossary the team references. The highest-risk processes captured as checklists. New files following a naming convention. A named owner for each artefact and a quarterly review on the calendar. Direction of travel, with a rhythm that survives the quarter after the sprint ends.
The data will still have errors. The documented processes will reveal gaps when somebody new follows them. The naming convention will land at 75 to 85 per cent adoption rather than 100. The glossary will need terms added and refined. None of that is failure. The business has moved from a state where the next AI tool will stall at the data layer to a state where it will not.
The next move is the AI use case the firm has already named, sales forecasting, customer segmentation, invoice processing, a knowledge assistant. Whichever it is, the implementation no longer spends its first month cleaning data that should have been clean. That is the practical case for the sprint. Want a hand pacing this for your own firm? Book a conversation.



