KPMG has laid out a series of safeguards aimed at preventing AI agents from producing unpredictable or harmful outcomes, a move that aligns with growing federal attention to AI governance. The firm’s approach echoes the structural logic of the U.S. government’s primary framework for managing AI risk, raising a practical question for enterprises racing to deploy autonomous systems: what does “keeping AI on the rails” actually require?
Why AI Agents Need Guardrails Now
AI agents differ from traditional software in a critical way. Rather than executing fixed instructions, they make decisions, take actions, and adapt their behavior based on changing inputs. That autonomy creates real value in areas like financial analysis, customer service, and healthcare triage. But it also introduces failure modes that static rule-based systems never faced. An agent that misinterprets a prompt, hallucinates data, or optimizes for the wrong objective can cause cascading errors before any human notices.
KPMG’s safeguard recommendations target exactly this gap. The firm has outlined operational controls centered on continuous monitoring, human oversight checkpoints, and structured accountability across the AI lifecycle. These are not abstract principles. They map closely to the four core functions described in the federal government’s own risk management guidance: govern, map, measure, and manage.
NIST’s Four-Function Framework as a Benchmark
The U.S. National Institute of Standards and Technology has published the AI risk management framework, which serves as the primary federal reference for organizations trying to systematize AI safeguards. It organizes risk mitigation into four functions. “Govern” establishes policies and accountability structures. “Map” identifies the context and conditions under which an AI system operates. “Measure” calls for quantitative and qualitative evaluation of risks. “Manage” focuses on prioritizing and acting on those risks once identified.
KPMG’s safeguards can be evaluated through this lens. The firm’s emphasis on human oversight checkpoints, for instance, fits squarely within the “govern” function, which calls on organizations to integrate accountability at every stage of AI deployment. Its focus on continuous monitoring aligns with “measure” and “manage,” where ongoing evaluation and corrective action prevent small errors from compounding into systemic failures.
That alignment is not coincidental. NIST designed AI RMF 1.0 to be voluntary and sector-agnostic, meaning consulting firms, banks, hospitals, and tech companies can all use it as a shared reference point. KPMG’s decision to build its safeguard recommendations around a compatible structure signals that the firm expects this framework to become a de facto standard, even without a legal mandate.
What KPMG’s Controls Actually Look Like
The specific safeguards KPMG has detailed go beyond generic calls for “responsible AI.” They include structured intervention points where humans review agent decisions before they are executed, particularly in high-stakes domains like lending, diagnostics, and regulatory compliance. They also include logging and audit trail requirements so that when an agent does produce an unexpected output, teams can trace the reasoning chain backward to identify where things went wrong.
This matters because most AI failures are not dramatic. They are quiet. An agent that subtly drifts in its risk scoring, or that begins favoring one data pattern over another without flagging the shift, can operate for weeks before the downstream effects become visible. KPMG’s approach treats this kind of silent drift as the primary threat, not the Hollywood scenario of a rogue system acting against its operators.
The firm also stresses the importance of defining clear boundaries for agent autonomy before deployment, not after. That means specifying which decisions an agent can make independently, which require human approval, and which are entirely off-limits. This pre-deployment scoping exercise is one of the most practical steps any organization can take, and it maps directly to the “map” function in NIST’s framework, which asks teams to identify the conditions and constraints under which their AI systems will operate.
On the technical side, KPMG’s recommendations typically call for layered controls rather than a single safeguard. Model validation, adversarial testing, red-teaming exercises, and scenario simulations are paired with process checks such as documented sign-offs and segregation of duties. The idea is that if one layer fails (say, a test set misses a particular bias pattern), another layer, such as human review of edge cases, can still catch the issue before it reaches production scale.
Where the Federal Framework Falls Short
NIST’s AI RMF 1.0 provides an authoritative structure, but it has limitations that KPMG’s work implicitly highlights. The framework is deliberately general. It does not prescribe specific technical controls, testing thresholds, or enforcement mechanisms. That flexibility is a strength for broad adoption, but it leaves a gap when organizations need concrete implementation guidance.
KPMG’s safeguards attempt to fill that gap by translating NIST’s abstract functions into operational checklists and decision trees. The question is whether this translation holds up under real-world pressure. A consulting firm’s recommended controls are only as strong as the client’s willingness to fund, staff, and maintain them. And unlike a regulatory requirement, a voluntary framework carries no penalty for non-compliance.
This is the central tension in AI governance right now. The federal government has provided a broad security and standards resource hub through NIST and its Computer Security Resource Center, but it has stopped short of mandating specific practices for AI agents. That leaves firms like KPMG in the role of translating government guidance into enterprise-ready controls, a role that carries influence but no enforcement authority.
In practice, this means that two organizations can both claim to be “aligned” with NIST while implementing radically different levels of rigor. One might invest in continuous monitoring, third-party audits, and robust incident response plans. Another might simply document a high-level policy and consider the box checked. Without clearer expectations or oversight, the label of compliance risks becoming more about optics than substance.
The Real Risk of Voluntary Compliance
Most coverage of AI safeguards treats them as a technical problem: build better guardrails, run more tests, add more monitoring. But the harder challenge is organizational. Companies deploying AI agents face constant pressure to move fast, cut costs, and demonstrate returns on their AI investments. Safeguards that slow deployment or add overhead are easy to deprioritize, especially when competitors are moving without them.
KPMG’s recommendations implicitly acknowledge this dynamic by framing safeguards not as constraints but as risk reduction measures that protect long-term value. The argument is that an AI agent that causes a regulatory violation, a biased lending decision, or a medical error will cost far more to remediate than the upfront investment in proper controls. That framing is sound, but it depends on executives making long-term calculations in a market that rewards short-term speed.
The absence of binding federal regulation makes this calculus even harder. Without a legal floor for AI safety practices, the range of implementation quality across industries will remain wide. Some firms will adopt KPMG-style controls rigorously. Others will treat them as aspirational guidance, implementing only the easiest or most visible pieces. Over time, that divergence could create a two-tier landscape in which well-governed AI systems coexist with minimally controlled agents operating in adjacent markets.
For now, the most realistic path forward is a hybrid model. Organizations can use NIST’s framework as the structural backbone, adopt detailed safeguard sets like KPMG’s as operational playbooks, and then layer on their own domain-specific controls. Boards and senior leaders will need to treat AI risk as a standing agenda item rather than a one-time project, with clear metrics for performance, safety, and compliance.
If AI agents are going to take on more decision-making authority inside critical systems, the cost of weak governance will only rise. Voluntary frameworks and consulting-led safeguards can move the market toward better practices, but they are not a substitute for sustained oversight and a willingness to slow down when the risks demand it. The organizations that internalize that lesson early are likely to be the ones still benefiting from AI automation when the next wave of failures forces a broader reckoning.
More from Morning Overview
*This article was researched with the help of AI, with human editors creating the final content.