Strategy 10 min read

Transparency Theater in AI Governance: When Safety Is a Performance

J

Jared Clark

March 21, 2026

There is a particular kind of document that has become almost obligatory in the AI industry. It has a title like "Our Commitment to Responsible AI" or "Principles for Safe and Beneficial AI." It features carefully chosen language about fairness, accountability, and human oversight. It is prominently linked from the company's homepage, cited in congressional testimony, and handed to journalists as evidence of good faith.

And then, in many cases, it changes almost nothing about how the company actually builds or deploys its systems.

This is transparency theater — and it has become one of the most consequential problems in AI governance today.


What Is Transparency Theater?

Transparency theater is the practice of publishing safety frameworks, ethical principles, and governance commitments that are structurally designed to appear meaningful while remaining too vague, too voluntary, or too unverifiable to constrain real behavior.

The term borrows deliberately from "security theater" — the concept popularized by Bruce Schneier to describe security measures that make people feel safe without actually making them safer. The airport shoe removal ritual is the canonical example: visible, inconvenient, and largely ineffective.

AI governance has developed its own version of this dynamic. The visibility is high. The inconvenience to companies is low. And the actual safety impact is, in many cases, difficult to verify at all.

This isn't a fringe critique. In 2023, the Stanford Center for Research on Foundation Models found that leading AI model developers scored an average of just 40 out of 100 on its Foundation Model Transparency Index — meaning that even the industry's most prominent players disclosed less than half of the information needed to evaluate the safety and societal impact of their models. That's not a transparency problem. That's a transparency performance.


The Five Hallmarks of Transparency Theater

Not every AI safety document is theater. But certain structural features reliably signal that a commitment is designed for optics rather than accountability. Here are the five patterns I've identified most consistently.

1. Principles Without Processes

The most common form of transparency theater is the publication of high-level ethical principles with no corresponding operational processes. Declaring that an AI system will be "fair" or "human-centered" means nothing without specifying: Who tests for fairness? By what metric? At what threshold does a system fail? Who has the authority to halt deployment?

When these questions go unanswered, principles function as a moral alibi rather than a governance mechanism.

2. Accountability Without Consequences

Real governance has teeth. Transparency theater doesn't. A telling sign is an AI ethics board or safety committee that has no power to veto, delay, or modify a product launch. Several high-profile ethics boards — including Google's short-lived Advanced Technology External Advisory Council, disbanded in 2019 after just ten days — have demonstrated that the appearance of oversight can be constructed and dismantled faster than the systems being "governed."

As of 2024, a study by researchers at the AI Now Institute found that fewer than 20% of corporate AI ethics policies included any enforcement mechanism or described consequences for non-compliance. The rest were aspirational statements.

3. Commitments That Are Not Falsifiable

Good governance produces claims that can be proven wrong. Transparency theater produces claims that cannot. Compare these two statements:

Statement Type Example
Falsifiable Commitment "We will conduct red-team testing for at least 90 days before any frontier model release and publish a summary report."
Unfalsifiable Claim "We are committed to the responsible development of AI."
Falsifiable Commitment "Our models will be evaluated against the NIST AI RMF prior to deployment, and results will be disclosed to regulators."
Unfalsifiable Claim "We believe AI should benefit humanity."
Falsifiable Commitment "We will not deploy models that score above a defined harm threshold on our internal benchmarks."
Unfalsifiable Claim "We take safety seriously."

The ratio of unfalsifiable claims to falsifiable commitments in a governance document is one of the most reliable indicators of whether it is substance or performance.

4. Selective Transparency

Some companies publish impressive amounts of information about their AI systems — but carefully curate what that information covers. They may release detailed technical papers about model architecture while disclosing nothing about training data, safety incidents, or deployment decisions. They publish bias evaluations on benchmark datasets while staying silent on real-world failure rates.

This is not transparency. It is the management of disclosure — using openness in low-stakes areas to create the impression of openness overall.

5. Governance That Lags Deployment

Perhaps the most structurally revealing pattern is the timing of governance activity. In genuine safety cultures, governance precedes or accompanies deployment. In transparency theater, governance follows it — often by months or years, and often only in response to external pressure or a high-profile incident.

A 2023 report from the Center for AI Safety noted that most AI companies lack pre-deployment evaluation standards that would be recognized as rigorous by independent safety researchers. Governance documents that are published after a system is already in wide use are not safety frameworks. They are reputation management.


Why Transparency Theater Flourishes

Understanding why this dynamic exists is essential to changing it. Transparency theater isn't simply cynical. It emerges from a specific set of structural incentives.

The Regulatory Vacuum

Where regulation is absent or ambiguous, voluntary commitments fill the space — and voluntary commitments without verification mechanisms tend toward the aspirational. In the United States, comprehensive federal AI regulation remains unenacted as of 2025, leaving companies free to define what "responsible AI" means for themselves. The EU AI Act represents the most substantive regulatory framework to date, but its full enforcement provisions don't take effect until 2026–2027.

In this environment, publishing a safety framework costs almost nothing and signals virtue without requiring verification. The incentive is obvious.

The Competitive Dynamics of Frontier AI

The frontier AI industry is characterized by intense competition and enormous first-mover advantages. Being first to deploy a capable model has significant commercial and strategic value. This creates structural pressure to treat safety governance as a cost to be minimized rather than a commitment to be honored.

When a company's most direct competitors are also publishing aspirational safety documents without meaningful accountability, the reputational pressure to do more is limited. The floor becomes the ceiling.

The Technicality Problem

There's also a genuine epistemic challenge: many AI safety commitments are made about risks that are difficult to evaluate even in good faith. Emergent behaviors in large language models, dual-use potential, and long-horizon risks are hard to measure. Some companies use this genuine uncertainty as cover for commitments that are deliberately vague.

The result is a space where it is difficult to distinguish between "we are honestly uncertain about how to evaluate this risk" and "we are using uncertainty as a rhetorical shield." Both look similar from the outside.


How to Read an AI Governance Document Critically

If you are an enterprise evaluating an AI vendor, a policymaker assessing industry self-regulation, a journalist covering AI, or simply a thoughtful observer, you need a framework for distinguishing substance from performance. Here's how I approach it.

Ask: What Would Prove This Wrong?

Apply the falsifiability test to every major commitment in the document. If a commitment cannot, even in principle, be violated, it is not a commitment. It is a sentiment.

Ask: Who Has the Power to Say No?

Identify the governance bodies named in the document. Then determine: Does this body have the authority to halt or modify a deployment? Are its decisions binding? Is it staffed by people with genuine independence from the product and revenue teams? If the answers are no, no, and no — the body exists for optics.

Ask: What Is Being Disclosed vs. What Is Being Hidden?

Map the disclosures against the full landscape of information that would be relevant to evaluating AI safety:

Information Category Disclosed?
Model architecture and training approach Often yes
Training data composition and sourcing Rarely
Red-team and adversarial testing results Occasionally
Real-world incident reports Almost never
Internal safety thresholds and benchmarks Almost never
Deployment decision criteria Rarely
Governance body meeting minutes or decisions Almost never

The pattern of disclosure is itself informative. Transparency that is systematically absent in the highest-stakes categories is not transparency — it is strategic omission.

Ask: What Changed After This Was Published?

The most powerful test of a governance document is behavioral: Did anything change after it was published? Were there product modifications? Deployment delays? Decisions not to release a capability? If the publication of a safety framework correlates with no observable changes in behavior, it was likely designed to manage perception rather than behavior.


What Genuine AI Governance Actually Looks Like

The case for distinguishing theater from substance isn't purely critical — it also points toward what real governance requires. Across the organizations and frameworks I find most credible, several features recur consistently.

Pre-deployment evaluation standards with defined pass/fail criteria. Not "we will evaluate safety" but "we will not deploy until these specific benchmarks are met, and here is how we will verify them."

Independent red-teaming with published summaries. The commitment to adversarial testing means little if the results are internal. Summaries — even redacted ones — that describe what was tested, what was found, and how it was addressed demonstrate genuine engagement.

Governance bodies with real authority. This means the power to delay or halt deployment, staffed by people who are not directly subordinate to product or commercial leadership, with documented decision records.

Incident reporting and post-mortems. Genuine safety cultures treat failures as learning opportunities that are documented and shared. An organization that has never published an AI incident report — despite deploying systems at scale — is telling you something important about its relationship with accountability.

Time-bound, specific commitments. "By Q3 2025, we will publish evaluation results for all externally deployed models" is a real commitment. "We will continue to improve our safety practices" is not.

The gap between these standards and current industry practice is substantial. A 2024 analysis of the 10 largest AI model developers found that none fully met all five criteria that independent safety researchers consider minimum requirements for credible pre-deployment governance. That gap is where the work of genuine reform is located.


The Stakes of Getting This Wrong

Transparency theater is not merely a public relations problem. It has material consequences.

When policymakers rely on voluntary safety commitments that are structurally designed to be unverifiable, they are building regulatory frameworks on sand. The EU AI Act, the White House Executive Order on AI (2023), and various international governance initiatives have all incorporated elements of voluntary industry commitment. If those commitments are theater, the governance frameworks built on them are too.

When enterprises procure AI systems based on vendor safety documentation, they are making risk management decisions on the basis of that documentation. If the documentation is theater, the risk management is theater too — and the liability eventually lands somewhere real.

And when the public is asked to accept the deployment of consequential AI systems on the basis that companies have made safety commitments, the erosion of those commitments corrodes the broader social license on which AI development depends. Transparency theater, practiced at scale and over time, is a mechanism for destroying trust — not building it.

The good news is that the analytical tools for distinguishing theater from substance are accessible. Falsifiability. Independent authority. Disclosure patterns. Behavioral change. These aren't technical criteria that require AI expertise. They are the same criteria we apply to accountability in any other high-stakes domain.

The harder challenge is political: creating the incentive structures — regulatory, reputational, and market-based — that make genuine governance more valuable than performed governance. That requires pressure from regulators, enterprises, journalists, and the public alike.


The Question Worth Asking

The next time an AI company publishes a safety framework, a responsible development charter, or a set of AI principles — before you read the content, ask a simpler question: What would have to happen for this company to acknowledge that it violated its own commitments?

If you can't answer that question from the document, the document isn't doing governance. It's doing something else.

And that something else has a name.


For more on how organizations can build genuine AI readiness rather than performed readiness, explore how AI is transforming institutional accountability and what real AI preparedness looks like in practice on Prepare for AI.


Last updated: 2026-03-21

J

Jared Clark

Founder, Prepare for AI

Jared Clark is the founder of Prepare for AI, a thought leadership platform exploring how AI transforms institutions, work, and society.