Skip to main content
FIELD REPORT · AI

AI Adoption Playbook for Finance Teams: From Pilot to Production

A concrete guide for CFOs and FP&A leaders on the six AI use cases that actually work in finance, plus vendor selection, ROI math, and the SOX implications most teams miss.

PUBLISHED
May 6, 2026
READ TIME
13 MIN
AUTHOR
ONE FREQUENCY

Finance teams sit on more structured data than almost any function in the enterprise, yet most CFOs we work with are running AI pilots that produce slide decks instead of cycle-time reductions. The pattern is predictable: a vendor demo, a steering committee, a six-week pilot, and a conclusion that "the technology is promising but not yet ready." The technology is ready. The deployment discipline usually is not.

This playbook covers the six AI use cases that have produced measurable value for finance teams in the last 24 months, what vendor and build options exist for each, what the audit and SOX implications are, and where the landmines sit. If you are a CFO, controller, or FP&A leader weighing where to spend your 2026 AI budget, this is the operating picture you need before signing a single SOW.

The six finance use cases that actually return value

Most "AI for finance" vendor pitches blur together. Strip them down and you get six distinct workflows where the math works. The rest are still science projects.

| Use case | Time to value | Typical ROI year 1 | Risk profile | |----------|---------------|--------------------|--------------| | Variance analysis & commentary | 6-10 weeks | 30-50% FP&A cycle time reduction | Low | | Contract review & abstraction | 4-8 weeks | 60-80% review time reduction | Medium | | Forecasting & scenario modeling | 12-20 weeks | 10-25% forecast accuracy improvement | Medium | | Expense audit & policy enforcement | 4-6 weeks | 3-7% T&E spend recovery | Low | | Invoice processing (AP automation) | 8-16 weeks | 50-70% touchless invoice rate | Medium | | Board pack & narrative drafting | 2-4 weeks | 40-60% drafting time reduction | Low |

These numbers come from production deployments, not vendor brochures. The variance is mostly explained by data quality, not the AI itself.

Variance analysis and FP&A commentary

The unglamorous reality of FP&A is that analysts spend the bulk of close week copying numbers from BI tools into PowerPoint, then writing two-sentence explanations of why the number moved. Large language models do this competently when given the actual underlying data and a clear prompt structure.

Microsoft 365 Copilot for Finance plugs directly into Excel and Dynamics 365 and will draft variance commentary against your actuals-vs-budget pivots. Anaplan AI does the same thing inside Anaplan models. For teams on NetSuite or SAP, the build-it-yourself path using the Anthropic or OpenAI API against a structured data warehouse is straightforward — 4 to 6 weeks for a competent data engineering team.

The trap: never let the model generate numbers. It should explain numbers that come from your system of record. The prompt pattern is "given these actual figures, write commentary." Never "what was the revenue this quarter."

Contract review and abstraction

Procurement, legal, and revenue ops all consume contracts. AI does first-pass extraction extremely well — payment terms, auto-renewal clauses, termination triggers, MFN provisions, indemnity caps, governing law. Ironclad, LinkSquares, and SirionLabs have built-in AI extraction. The newer entrants (Harvey, Spellbook) target legal review more than procurement abstraction.

Expect 60-80% reduction in review time for standard agreements. Non-standard agreements still need a human pass, but the human starts from a populated abstract instead of a blank page.

Forecasting and scenario modeling

This is where the gap between hype and reality is widest. AI does not magically improve your forecast. What it does is let you run more scenarios faster and surface drivers you would not have looked at. Workday Adaptive Planning, Anaplan, and Pigment all have AI features that auto-generate scenarios and identify outlier inputs.

The 10-25% accuracy improvement happens when AI is used to identify previously-ignored drivers (lead indicators, external signals) and to widen the scenario set, not when AI is asked to predict the future on its own.

Expense audit and policy enforcement

Brex Empower, Ramp Intelligence, and SAP Concur all now use AI to flag out-of-policy expenses, duplicate submissions, and patterns suggestive of fraud. The 3-7% spend recovery is real, especially in companies that previously sampled expense reports rather than reviewing all of them.

Invoice processing

AP automation is the most mature AI use case in finance. Tipalti, AppZen, Stampli, and Vic.ai all do OCR + classification + GL coding + approval routing with high accuracy. A 50-70% touchless rate means more than half your invoices flow from receipt to payment without human touch. The remaining 30-50% are exceptions that still need human judgment.

The trap: vendors will quote you the touchless rate from their best customer. Yours will be lower until your master data (vendors, GL accounts, cost centers) is clean.

Board pack and narrative drafting

Drafting the CEO letter, the MD&A section of the 10-Q, or the board narrative is high-leverage work for finance leadership. Models trained on your prior filings and using your current financials will produce competent first drafts in minutes. The human edit is still essential, but the draft saves hours per cycle.

Build vs. buy: a decision rule

For each of the six use cases, you can buy a finance-specific SaaS, build on a horizontal platform (Microsoft Copilot, Google Duet, Anthropic Claude), or roll your own on a foundation model API.

The simple rule: buy if your data already lives in the vendor's system. Build if you need to span multiple systems or if your workflow is unusual. Roll your own only when no vendor solution fits and your engineering team has the discipline to maintain it.

If you run NetSuite, your AP automation should probably plug into NetSuite, not be a generic best-of-breed. If you run SAP S/4HANA, Joule is the path of least resistance. If your data is split across five systems and you are mid-ERP migration, a horizontal layer (Anthropic Claude with custom tooling, or a Snowflake-native AI app) usually beats locking into one ERP vendor's roadmap.

The SOX and audit trail problem

Here is the part most pilots ignore until quarter close. If AI is touching any process that flows into the general ledger, SOX 404 applies. That means you need:

  1. Documented controls describing what the AI does, what data it consumes, and what human review occurs.
  2. Reproducibility. If an auditor asks why a journal entry was made, you need to reproduce the AI's reasoning. Most LLM outputs are non-deterministic. Set temperature to 0 where possible and log the prompt, model version, and output for every transaction.
  3. Access controls consistent with the rest of your financial systems. The AI should not have broader read access than the human it replaces.
  4. Change management. Model version changes are software changes. Treat them like any other ITGC change.
  5. Segregation of duties. The same model should not be both proposing and approving a journal entry. Most teams forget this and end up with a finding.

Your external auditor probably does not have a formal AI audit program yet. Get ahead of them. Document everything before they ask.

The hallucination problem in finance specifically

A model that hallucinates a fact in a marketing draft is annoying. A model that hallucinates a number in a variance commentary is a restatement risk.

The mitigation is architectural, not behavioral. Never ask a language model to produce a number that originates with the model. The model should retrieve numbers from your system of record, transform them deterministically, and explain them. Use retrieval-augmented generation (RAG) patterns where the model cites the source row for every number it surfaces. If your vendor cannot show you the retrieval architecture, you do not have the architecture you need.

A finance-specific AI risk matrix

| Risk | Likelihood | Impact | Mitigation | |------|-----------|--------|------------| | Hallucinated number in external report | Medium | Severe | Deterministic retrieval, no model-generated numerics | | Audit trail gap | High (default) | High | Prompt + output logging, model version pinning | | Unauthorized PII or MNPI exposure | Medium | Severe | Data classification before ingestion, region-pinned models | | Vendor data training on your data | Medium | High | Enterprise contract with zero-retention guarantee | | SOX deficiency on AI-touched control | High (default) | High | Document controls before deployment, not after | | Forecast over-reliance | Medium | Medium | Maintain human-led baseline forecast in parallel for 2 cycles |

A 90-day deployment arc

Week 1-2: Pick one use case. One. Resist the urge to do three.

Week 3-4: Data readiness. Pull a representative dataset, classify it, confirm retention and access policies, and prove the vendor or platform can meet them.

Week 5-6: Build the technical integration. For a SaaS, this is mostly configuration. For a build, this is the bulk of the work.

Week 7-8: Pilot with a single business unit or controller's team. Real workflows, not synthetic data.

Week 9-10: Measure against the baseline you captured in week 1. Cycle time, accuracy, exception rate.

Week 11-12: Go/no-go decision. Document the SOX implications, sign off with internal audit, then scale.

If you are weighing this against a broader AI program, our AI implementation roadmap for the enterprise covers how finance fits into the larger sequencing decisions, and the AI governance framework template gives you a starting point for the policy work that needs to happen in parallel.

Data readiness: the prerequisite most teams skip

Every finance AI deployment we have triaged had the same root cause when it stalled: the data was not ready. Chart of accounts inconsistencies across entities. Vendor master records with duplicates and typos. Cost centers that mean different things in different business units. Currency conversions that happen at three different layers.

The model amplifies whatever it consumes. Bad data produces confidently wrong AI output, which is more dangerous than confidently wrong human output because the human knew their limits.

A 30-day data readiness assessment before a finance AI pilot should cover:

  1. Chart of accounts hygiene. How many active GL accounts? How many should be inactive? Are there parallel hierarchies for management vs. statutory reporting?
  2. Vendor master deduplication. Run a fuzzy match across vendor names. The duplicates will surprise you.
  3. Cost center and project taxonomy. Are they consistently used across business units? Across systems?
  4. Reconciliation between source systems. Does the ERP agree with the consolidation tool agree with the BI warehouse? On the same day?
  5. Currency and FX handling. Where is the rate sourced? How often is it updated? What rates are used for translation vs. transaction?
  6. Period close discipline. Are sub-ledgers closing on the same cadence as the GL?

If any of these are weak, fix them first. The fix produces value on its own and dramatically improves the AI pilot odds.

How the new finance AI tooling integrates with existing stack

Pilot success is not just about the AI vendor. It is about how that vendor fits the rest of your stack. A short integration matrix to think through:

| Layer | Common systems | AI integration questions | |-------|----------------|--------------------------| | ERP | NetSuite, SAP S/4HANA, Oracle Fusion, MS Dynamics 365 | Native AI features vs. external? API rate limits? | | EPM/CPM | Anaplan, Workday Adaptive, Pigment, OneStream | Native AI features? Model-as-input or output? | | Procurement | Coupa, Ariba, Ivalua | Spend visibility AI? Contract AI? | | AP automation | Tipalti, Bill.com, AppZen, Stampli, Vic.ai | Touchless rate? GL coding accuracy? | | T&E | Brex, Ramp, Concur, Navan | Real-time policy enforcement? Fraud detection? | | Close & consolidation | BlackLine, FloQast, Trintech | Reconciliation AI? Variance detection? | | Tax | Vertex, Avalara, Sovos | Determination accuracy? Audit defense? | | Reporting | Workiva, Tableau, Power BI | Narrative generation? Chart commentary? |

A finance AI strategy that picks vendors layer by layer typically ends up with seven or eight overlapping AI products. Pick the strategic three or four and accept that some layers will not have AI for a year or two.

A pilot governance model that survives audit

The pilot needs the same governance scaffolding as production. Cheaper to build it once, in pilot, and then scale.

Minimum governance components:

  1. AI inventory. A registered list of every AI deployment in finance, what it does, who owns it, what data it touches, what controls apply.
  2. Pre-deployment checklist. A documented gate that every pilot passes before production. Includes legal, privacy, audit, IT security, and finance leadership sign-off.
  3. Model card. A one-page summary of the model, training data, intended use, known limitations, and review cadence.
  4. Incident log. When something goes wrong (hallucination, downtime, wrong number), it gets logged with root cause and remediation.
  5. Quarterly review. Finance leadership reviews the AI inventory, incident log, and ROI metrics on a fixed cadence.

Skip the governance and your CFO is one bad headline away from killing the program. Build it and you have something you can show the board.

Common pitfalls we see

  1. The CFO buys five tools at once. Pick one, prove it, then add. You cannot change-manage five vendors simultaneously.
  2. No baseline measurement. If you do not know the current cycle time, you cannot prove improvement.
  3. No human in the loop on judgment calls. Models are good at synthesis, not at judgment. Keep humans on the close, on policy exceptions, on anything that goes to the board.
  4. Audit gets surprised in week 12. Bring internal audit in week 1.
  5. The pilot succeeds and then dies. Without an explicit production owner and operating budget line, pilots evaporate when the sponsor moves on.
  6. No model version pinning. The vendor upgrades the model and your variance commentary changes tone overnight, breaking your audit trail.
  7. Treating AI output as authoritative. The model said it; therefore it is true. This is how restatements happen. Build the verification step.
  8. No off-boarding plan when a vendor is swapped. Data extracted; model weights retained somewhere; nobody knows what happened to the training data you gave them.

Next steps

If you are running a finance AI pilot now and not sure whether it will survive an auditor, or you are at the use-case selection stage and trying to avoid buying the wrong five tools, this is the work we do. We help finance leaders sequence the right use cases, write the SOX-ready controls, and stand up the technical integrations. Reach out when you are ready to move past pilots that produce decks.

View All Insights
NEXT STEP

Ready to ship the next outcome?

One Frequency Consulting brings 25+ years of technology leadership and military discipline to every engagement. First call is operator-grade scoping — sixty minutes, no charge.