Skip to content
← Back to Blog

The Guardrails Library Problem: Why Runtime Filters Are Not a Substitute for Pre-Certification

The Guardrails Library Problem: Why Runtime Filters Are Not a Substitute for Pre-Certification

The AI safety tooling market has consolidated around a convenient fiction: that scanning an agent's output after it runs is equivalent to certifying what the agent will do before it runs. It isn't. And as enterprises start attaching real business consequences to AI agent behavior—automated purchasing, customer communications, financial decisions—the gap between these two models stops being academic.

What the Market Actually Built

The guardrails library ecosystem grew up solving a legitimate problem. Early LLM deployments produced outputs that were unpredictable, sometimes harmful, and difficult to audit. Libraries emerged to filter that output: block profanity, detect PII, flag toxic language, enforce response schemas. Gartner estimated the AI trust, risk, and security management (TRiSM) market at roughly $1.5 billion in 2023, with significant growth projected through 2026. The dominant pattern across that market is post-execution intervention.

This made sense when AI agents were stateless question-answering systems. You ask, the model responds, a filter checks the response. The blast radius of a bad output is a bad response.

That architecture has been carried forward, largely unchanged, into a fundamentally different deployment environment: autonomous agents that take actions, call external APIs, modify databases, and operate across multi-step reasoning chains. The filter is still sitting at the end of the pipeline. The actions have already happened.

The Execution Gap Is the Actual Risk Surface

Here's the concrete problem. A guardrails library operating at output layer can tell you that the text the agent returned doesn't contain a credit card number. It cannot tell you whether the agent called a payment API mid-execution with one. Runtime filters intercept artifacts. They don't intercept behaviors.

Research on agentic systems published through 2023 and 2024 consistently identifies the intermediate steps of agent execution—tool calls, sub-agent spawning, memory writes—as the primary risk surface, not the final natural language output. A study from Stanford's HAI found that agent failures in multi-step tasks disproportionately occurred at action boundaries, not at response generation. Post-execution filters are structurally blind to this layer.

The compliance implication is direct. If your auditor asks whether your agent was certified not to access customer financial records during a specific workflow, "our output filter didn't flag anything in the final response" is not an answer to that question. It's a different question answered confidently.

Pre-Execution Certification Is a Different Contract

Pre-execution safety certification operates on a different premise: you certify the agent's behavior envelope before it runs, not the content it produces after it runs. Concretely, this means defining the specific inputs, tool access, decision logic, and output constraints that constitute compliant operation—then generating a verifiable certificate, cryptographically bound via SHA-256 hashing, that attests to that configuration at a specific point in time.

The certificate answers a different question than the filter does. The filter answers: "Did this output look acceptable?" The certificate answers: "Was this agent operating within a verified behavioral specification when it ran?"

This matters enormously for deterministic workflows. If an agent, given the same inputs, must produce the same action sequence—and that sequence has been certified—you have a reproducible compliance claim. Any deviation from certified behavior is detectable as a configuration drift event, not discovered post-hoc when a customer complaint surfaces.

This model also works across agent frameworks without modification. Whether teams are building with custom REST-based orchestration or using any existing framework, the certification layer sits at the workflow definition level, not the output level. It's framework-agnostic by design.

The False Compliance Signal Is the Real Liability

The market risk practitioners should be tracking isn't that guardrails libraries are bad tools. They're useful for what they were designed for. The risk is that organizations are checking a compliance box with a tool that was never designed to satisfy that compliance requirement.

An enterprise deploying an autonomous procurement agent, a claims-processing agent, or a patient-scheduling agent isn't primarily worried about whether the agent's summary text sounds professional. They're worried about whether the agent's actions were within authorized parameters. Runtime filters don't certify that. They filter text.

As regulatory frameworks around AI systems mature—the EU AI Act's requirements for high-risk system documentation being the most immediate example—the distinction between "we filtered outputs" and "we certified pre-execution behavior" will become a legal distinction, not just a technical one.

Organizations building on agent infrastructure now need to make that distinction clearly, in their architecture decisions and in their compliance documentation. The guardrails library handles one problem. Agent certification handles a different and, for agentic deployments, more consequential one. Conflating them doesn't create safety. It creates the appearance of safety, which is worse.

Get your free API key at api.qaesubstrate.com

Get API Key →