Finance analysts, legal professionals, marketing teams, and operations managers are the group most likely to be producing AI-assisted outputs that reach external parties: clients, regulators, partners, investors. They're also the group for whom getting AI wrong is most directly traceable to the organisation's legal and reputational exposure.
The figures that put this in context: AI hallucinations create liability in 17 to 34% of AI-assisted legal workflows, and enterprises report financial losses linked to hallucinations in up to 11% of AI deployments. Those figures aren't distributed evenly across the workforce. They concentrate in exactly the functions this article covers, where AI-generated outputs carry the organisation's name, inform decisions with material consequences, or both.
Generic AI literacy training doesn't address this. What these four functions need is training grounded in the specific failure modes AI produces in their work, the specific accountability structures they operate within, and the specific judgment calls their domain requires. This article builds on the broader role framework in AI Literacy by Role and goes deeper into each function.
Why Domain Expertise Is the Critical Variable
The training gap for knowledge workers isn't primarily about tools. It's about domain judgment applied to AI outputs. A general-purpose AI literacy course teaches employees that AI can be wrong. What it cannot teach is what wrong looks like in a financial model, a contract clause, a customer segment, or a supply chain forecast.
When courts sanction lawyers, they hold counsel responsible regardless of who selected the tool or how sophisticated the vendor's claims were. The same principle applies across all four functions: accountability attaches to the professional who acted on or approved the output, not to the model that produced it.
This means AI training for knowledge workers is inseparable from professional competency training. It's not an add-on. The training that produces a finance analyst who can catch an AI-generated error in a forecast is the same training that produces a competent finance analyst. It just needs to be explicitly designed for the AI-assisted workflow they're actually operating in. That's what each of the four function-specific sections below is built around.
Finance Teams — When a Wrong Number Is Irreversible
Finance is one of the highest-risk functions for AI-assisted error because the consequences of a wrong number are immediate, auditable, and often irreversible. AI systems can extract relevant data from complex financial documents with higher accuracy than manual review. But poor-quality training data can perpetuate discrimination in credit scoring and robo-advice. And many AI models used in finance are genuinely opaque: firms using tools like Zest AI, Upstart, or Moody's CRE for credit and risk assessment often can't interpret the outputs, identify root causes of errors, or defend decisions to regulators or customers.
Three specific failure modes concentrate most of the risk. First, AI-generated forecasts that present plausible figures with no indication of the assumptions underneath them — a Bloomberg or Refinitiv data feed processed through a generative layer gives a clean output that may embed assumptions the analyst never sees. Second, credit or risk scoring tools whose training data reflects historical bias, producing outputs that are statistically consistent but systematically skewed. Third, employees feeding sensitive data into public-facing tools like free-tier ChatGPT or Google Gemini and moving faster than internal controls can follow. It's now one of the primary compliance concerns for financial services firms in 2026.
A finance analyst receives an AI-generated quarterly revenue forecast. It's internally consistent and well-formatted. The analyst must identify three things it doesn't disclose: what the baseline assumptions are, whether recent market conditions are reflected in the training data, and what the confidence interval is around the central figure. They then write the three questions they'd ask before presenting this to a senior stakeholder.
Legal Teams — When Confidence Masks Fabrication
Legal is the function where AI hallucination creates the most direct professional liability. In late 2025, a major law firm was sanctioned for filing entirely fabricated ChatGPT-generated case citations in federal court — a case most in-house legal teams will be aware of. What's less well understood is that the same risk applies equally to in-house legal professionals using AI for contract analysis, regulatory research, and compliance assessments. You don't need to be at a law firm for this to become your problem.
Three properties of AI models make this particularly acute in legal contexts. First, hallucination: AI models confidently provide false information, and legal citation is exactly the output type where confidence can mask fabrication. Second, discriminatory outputs: AI-embedded legal tools may produce outputs that violate applicable laws, particularly in employment and contract contexts. Third, model drift: a contract analysis tool like Kira, Luminance, or Ironclad AI that passed a review in Q1 may behave differently by Q3, with no visible signal that anything has changed.
The practical training gap is that legal professionals using these tools don't have a structured habit for verifying outputs against source documents. They trust the summary because the summary looks authoritative. That's a habit problem, not a knowledge problem. Habit problems require practice, not explanation.
An in-house legal professional receives an AI-generated contract analysis flagging three risk areas in a supplier agreement. One flagged item contains a hallucinated clause reference that doesn't exist in the actual contract. One correctly identifies a genuine risk. One misses a material exclusion clause entirely. The task is to identify which is which and explain the professional consequence of acting on each without independent verification.
Accountability attaches to the professional who acted on the output, not to the model that produced it. Courts have consistently held counsel responsible for AI hallucinations regardless of which department selected the tool or how sophisticated the vendor's claims were. See deployer obligations guide for the regulatory dimension of this accountability.
Marketing Teams — When the Claim Is Already in the Contract
Marketing is the function where AI-generated errors most directly reach external audiences, and where accountability for those errors is least well understood. The liability concern isn't just reputational. AI-washing is actionable in cases where companies described AI capabilities they didn't actually have. That's an early sign that false AI claims may trigger securities liability, and it applies to marketing copy as much as to investor communications.
The more immediate day-to-day risk is simpler. Companies remain legally responsible for statements made through their sales and marketing processes, especially if the output was sent to a prospect, included in a proposal, or relied on during negotiations. AI does not remove that accountability. A Jasper or Copy.ai-generated campaign email that includes a performance benchmark the product has never actually achieved is a problem regardless of how that claim got into the email.
Three specific risks are most common. First, AI-generated copy that makes unverifiable or false product claims, particularly in B2B contexts where those claims end up in contracts. Second, AI audience segmentation built on biased or outdated data — a live concern for teams using platforms like Meta Advantage+ or Google Performance Max where the segmentation logic is largely opaque. Third, AI content tools generating outputs that incorporate third-party IP without attribution, a consistent issue with image generation tools and AI writing tools trained on scraped web content.
A marketing professional receives an AI-generated email campaign for a B2B software product. The email contains three specific claims: a performance benchmark, a security certification, and a customer success statistic. The professional must identify which can be used as-is, which requires verification before use, and which represents a legal exposure if published without independent sourcing. They then write the verification step required for each flagged claim.
Operations Teams — When the Error Has Already Compounded
Operations is the function where AI errors are most likely to be invisible until they've compounded. AI is embedded in supply chain forecasting, logistics optimisation, demand planning, and process automation. In most of these contexts the output isn't a document a human reads before acting on. It's a decision the system has already made.
This is a materially different failure profile from the hallucination risks that dominate conversation in legal and marketing. For operations professionals using platforms like Blue Yonder, o9 Solutions, or SAP Integrated Business Planning, the risks are more likely to be silent drift, biased optimisation, and cascading errors in automated pipelines. Mapping AI use cases and identifying where existing controls may not be designed for AI-driven failure modes has become a foundational step in operational risk management in 2026.
Three failure modes matter most for operations training. First, AI demand forecasts that embed assumptions from historical data that's no longer relevant — a tool trained on pre-2022 supply chain data has never seen the disruption patterns that have become normal since then. Second, supply chain AI tools that optimise for cost without surfacing the safety or ethical trade-offs that optimisation produces. Third, process automation tools that quietly remove human checkpoints, not through deliberate design but through accumulated workflow changes that nobody reviewed in aggregate.
An AI-powered inventory management system, using a platform like Relex or Llamasoft, has been recommending reduced safety stock levels for six consecutive weeks based on demand trend analysis. A weather event disrupts supply. The professional works backwards through the decision chain to identify where human oversight existed and was bypassed, where it existed but wasn't exercised on the right variable, and where it wasn't designed into the system at all. They then propose one specific change to the oversight design.
The Four Skills That Cut Across All Functions
Despite the different risk profiles above, four training needs apply to all four groups. All four are undertrained in standard AI literacy programmes. All four compound the function-specific risks above when they're missing.
Programme Design and Time Investment
Each function's training is designed as two modules: one covering function-specific risk scenarios, one covering the four cross-cutting skills applied to that function's context. Total time per function: 90 minutes.
| Function | Module 1 | Module 2 | Time |
|---|---|---|---|
| Finance | Domain-specific scenario analysis | Data classification + drift recognition | 90 minutes |
| Legal | Contract analysis critique | Citation verification + escalation | 90 minutes |
| Marketing | Claim verification exercise | IP and data hygiene in content | 90 minutes |
| Operations | Oversight gap analysis | Drift recognition + escalation | 90 minutes |
Each module produces a written deliverable: a classification, a set of verification steps, or a proposed governance change. That deliverable functions as both an assessment and a document the professional can use in their actual work. Programmes that skip the written output consistently produce lower rates of behaviour change, regardless of how well the scenario content lands in the session itself.
They're the ones whose training was specific enough to change what they do when they encounter an AI output in their actual work. Savia's role-specific AI learning paths are designed around exactly that standard.