I support your cause, I believe that whistleblowing is important.
I'm curious, how do you see the most efficient way to force AI companies to implement whistleblower protection policies? Do you have a concrete plan? E.g. legislative proposals.
Sure:
So for one, all of the companies listed above already claim to have internal whistleblowing channels, incl. retaliation protections. You can find sources for those claims in the campaign page if you're interested.
Legal provisions also already exist in the U.S. today that prevent companies from retaliating against individuals raising violations of the law - although not requiring a dedicated channel. Senator Grassley's proposed AI Whistleblower Protection Act would more concretely require AI companies to set up internal channels. We would indeed be happy to see more details in this legislation around guarantees e.g. 'independence' of a system should look like as well as transparency requirements (our level 1 and 2).
The issue even with mandating internal channels broadly is that a) many risks we might care about are currently still not covered under any law and b) that just claiming to have a policy/ system doesn't mean workers (at all) that there won't be retaliation - even if there's legal protections.
A) This means raising e.g. violations of an RSP would not be protected - unless the company explicitly includes them in a whistleblowing policy that is transferred into individually enforceable rights (i.e. that the individual can take a company to court over alleged retaliation for making a report) by a person covered under the policy.
B) If reports, for example, go straight to the executive board or managers involved in the misconduct, who then intimidate individuals speaking up so that they will never make use of their rights -> very bad but sadly not uncommon. Likewise if e.g. reports go to HR, but HR does not understand/ know how to deal with reports -> not good but not uncommon. Unfortunately there's also still many cases (in general, not AI specific) where retaliation is also more subtle or part of the system - e.g. a team member is 'moved around' instead of individuals responsible for misconduct. We will only know this if we get to "Level 2".
The EU e.g. mandates internal channels - retaliation still happens - both as systems may be poorly set up (Level 1) AND if it is not really 'lived' (Level 2).
So how do we get to companies improving protections in the status quo: We believe transparency on what current systems are and how they perform is the catalyst to improve systems. And if companies do not create that transparency (without compelling reason) - then we might also have an answer there about how reliable those systems are. As a note: This is not a 'silver bullet' solution but should be a strong improvement on the margin.
So how do we convince companies? This is from less to more 'pressure':
We of course also have thoughts on what an ideal whistleblowing policy (level 1) and public reporting on outcomes/ effectiveness/ improvements (level 2) should be - if you'd like an indication, take a look at the rating framework that FLI used in their AI Safety Index on the whistleblowing section, Appendix, Indicator "Whistleblowing Policy Quality Analysis": https://futureoflife.org/wp-content/uploads/2025/07/FLI-AI-Safety-Index-Report-Summer-2025.pdf
We'll publish a proper framework for evaluating systems in due time - but for new want to keep our focus on transparency.
Can I be useful for your work on whistleblowing? I'm an evals researcher, but I want to do some advocacy/lobbying/policy work as well.
Frontier AI companies currently lag global standards and best practices in creating adequate transparency around their internal whistleblowing systems: 5 out 6 companies in our target set do not even publish their whistleblowing policies.
That means we, the public, and employees in AI, are forced to 'trust' that companies will address concerns well internally.
This is far from good enough
...and why we, at the National Whistleblower Day Event in Washington DC yesterday, launched a campaign asking AI companies to publish their internal whistleblowing policies ("Level 1") and reports on their whistleblowing system performance, effectiveness, and outcomes ("Level 2").
We are very proud of the coalition we have the privilege of representing here - uniting most of the world's most prominent whistleblowing organizations and scholars with equally prominent AI counterparts.
See further below a full list of signatories or our campaign page.
You can find the actual campaign page, including evidence and sources, here: https://publishyourpolicies.org/
In this post I'll share the same message with a slightly altered 'storyline'.
I don't have to make the case here for why we should care about the way AI companies go about development and deployment of their frontier models - especially over the coming years.
Likewise, if you've seen righttowarn, you're likely aware of this line of reasoning: Many risks will only be visible to insiders. The current black-box nature of AI development means employees are often the first—and potentially only—people positioned to spot dangerous developments, misconduct, or safety shortcuts.
It therefore matters that AI companies build up the infrastructure required to address concerns raised today already and that we can enter a 'race to the top' on system quality as soon as possible.
Transparency on internal whistleblowing systems, allowing for public feedback and empowering employees to understand and compare protections is the mechanism to enter that 'race to the top' mechanism.
Important note 1: We are talking about company internal whistleblowing systems here (although they can extend arbitrarily far in terms of 'covered persons', e.g. to suppliers, customers, etc.). This does NOT diminish the importance of legal protections for AI whistleblowers or independent support offerings for insiders.
But the reality is (see below) that we expect the majority of risks to be flagged internally first. That means internal channels are critical and must not be neglected. If you like the 'swiss cheese mode' of risk management - we want to make sure protections are as strong as possible at every level.
Important note 2: Both in this post and our main post, we are not evaluating policy or system quality. We only talk about the degree of transparency provided.
Current and former AI employees have recognized that they are "among the few people who can hold [companies] accountable to the public." They've called for companies to "facilitate a verifiably anonymous process for current and former employees to raise risk-related concerns to the company's board, to regulators, and to an appropriate independent organization."
Research consistently shows that employees are often the first to recognize potential wrongdoing or risk of harm. In AI specifically, the technical complexity and proprietary nature of development means many risks are only visible to those with internal access.
Data from the SEC Whistleblower Program shows that three-quarters of award recipients initially attempted to address concerns within their organizations before seeking external remedies. Employees naturally try internal channels first, and we expect this to be no different in frontier AI companies:
This means that these systems must work reliably: When internal systems fail, we all lose. Companies miss opportunities to address problems early, employees face unnecessary risks, and the public remains unaware of safety issues until they potentially become crises.
Major AI companies have not published their whistleblowing policies. The recent Future of Life Institute AI Safety Index highlighted that Anthropic, Google DeepMind, xAI, and Mistral lack public whistleblowing policies, making neutral assessment impossible. They, likewise, call for the publication of policies.
Transparency of major AI companies’ whistleblowing systems*
*Please note that AIWI only evaluates the transparency of the policy and outcome reporting—not the content or quality of the underlying system, protections, culture, or past patterns of retaliation.
OpenAI is the sole exception—and they only published their policy following public pressure over their restrictive non-disparagement clauses. Even then, none of the major AI companies publish effectiveness metrics or outcome data.
This stands in stark contrast to other industries. Companies across sectors routinely publish whistleblowing policies—from AI-related organizations like ASML to industrial firms like Tata Steel to financial services companies. Many also publish regular effectiveness evaluations and outcome statistics.
Conversations with insiders also reveal gaps:
Employee Awareness: Interviews with current and former frontier AI company insiders show that many employees don't know, understand, or trust their companies' internal reporting systems. As one insider told us: "I'm not well-informed about our company's whistleblowing procedures (and it feels uncomfortable to inquire about them directly)."
Trust Deficit: AI employees suspect that making reports would be ineffective or could make their work lives more difficult. Another insider shared: "I anticipate that using official reporting channels would likely result in subtle, indirect consequences rather than overt retaliation like termination."
History of Retaliation: AI companies have attempted to suppress individuals voicing concerns (OpenAI's restrictive NDAs) and have faced cases around alleged wrongful termination for speaking up on research misconduct (Google).
We also have good reason to believe that multiple companies' internal whistleblowing policies are currently in violation of the EU Whistleblowing Directive. If you are interested: Happy to provide details via DM.
It might still be the case that certain systems are working relatively well today (at least for one of the organizations in the set we have an 'okay' impression based on conversations with individuals) - but the reality is that neither insiders nor we know.
Every insider we have spoken to to date supports the publication of whistleblowing policies. If you are an insider and you don't - please reach out and share your thoughts with us (or comment below).
Without published policies & outcome transparency, the public cannot assess whether internal systems actually protect employees who raise safety concerns.
Employees cannot compare protections across companies when making career decisions.
Policymakers cannot identify coverage gaps or craft appropriate regulations.
Companies benefit from improved systems through public feedback and heightened employee awareness. Empirical evidence shows that there is a strong 'business case' for improved speak up cultures and whistleblowing systems - from improved innovation to increased employee loyalty. This is why, for example, shareholder representatives have called on Google to improve its whistleblowing systems.
We are only calling for transparency: This should create no major workload for companies. If it does: Then maybe that means there were things to be improved upon).
Whistleblowing policies contain procedural frameworks and legal guarantees—not trade secrets or competitive advantages. There's no business case for secrecy, but substantial evidence for the benefits of transparency.
If companies truly care about developing a strong speak-up culture and protecting those who live it: Publish. Your. Policies.
We're calling on AI companies to meet two levels of transparency [this is an excerpt - see campaign page for details]:
Level 1: Policy Transparency (minimum baseline)
Level 2: Effectiveness Transparency (what companies should strive for)
Companies that take whistleblowing seriously should already gather this data for continuous improvement.
Publication is simply a matter of transparency.
This call is supported by a broad coalition of scholars, AI safety organizations, and whistleblowing advocacy groups:
Organizations:
Academic Signatories:
This campaign offers an opportunity for AI companies to demonstrate commitment to integrity cultures where flagging risks is a normal and expected responsibility.
We're not asking companies to reveal competitive secrets—we're asking them to show they're serious about the concern systems they claim to have. Transparency costs nothing but builds everything.
The stakes are too high for "trust us" to be enough. When AI companies publicly acknowledge existential risks, they must also demonstrate that employees can safely report concerns about those risks.
If you believe our call is sensible and you are...
Join the campaign: https://aiwi.org/publishyourpolicies/
Contact: For questions or to add your organization's support, reach out through the campaign website.
This campaign is led by The AI Whistleblower Initiative (AIWI, formerly OAISIS), an independent, nonpartisan, nonprofit organization supporting whistleblowers in AI.