TLDR: The EU’s Code of Practice (CoP) mandates AI companies to conduct state-of-the-art Risk Modelling. The current SoTA is not good. By creating risk models and improving methodology, we can enhance the quality of risk management performed by AI companies. This is a neglected area, and more people in AI Safety should be working on it. Work on Risk Modelling is urgent because the CoP is to be enforced starting in 9 months (Aug, 2, 2026).
Jan and Charbel contributed equally to this piece.
The AI Act and its CoP is currently the best regulation to prevent catastrophic AI risks. Many Frontier AI companies have signed the EU’s Code of Practice (CoP), and the AI Office, which enforces it, is staffed with competent personnel. The CoP requires companies to conduct risk management with respect to threats such as CBRN capabilities, Loss of Control, Cyber Offense, or Harmful Manipulation.
The code often refers to “state-of-the-art” and “appropriate” methods, instead of mandating specific Risk Management Practices. This opens a possibility for civil society and academics to raise the bar for AI companies by improving the state-of-the-art and building consensus on which strategies are appropriate.
In particular, the CoP requests a SoTA Safety and Security Framework (Commitment 1), risk modeling methods (Measure 3.3), model evaluations (Measure 3.2), and appropriate safety mitigations (Commitment 5).
Definition. From Campos et al 2025:
“Risk modeling is the systematic process of analyzing how identified risks could materialize into concrete harms. This involves creating detailed step-by-step scenarios of risk pathways that can be used to estimate probabilities and inform mitigation strategies.”
The other requested practices in the CoP seem to be less neglected or have less leverage:
In comparison, Risk Modeling enables:
And risk modeling is very neglected. There is little public work available on Risk Modelling for AI. Some notable work includes threat models on LLM-assisted spear phishing campaigns or the Replication of Rogue AI Agents. However, many essential risks don’t have a solid canonical risk model yet (see next section for examples), and most risk models can be harshly criticised.
This is an opportunity! There is a strong chance that conducting third-party research will influence what companies use in their reporting. If you improve the SoTA, in theory, companies should use your paper. Additionally, companies should consider publicly accessible information, which includes their risk models. METR's Autonomous Replication and Adaptation threat model demonstrates this: after building consensus around ARA as a severe risk, multiple frontier companies incorporated it into their RSPs.
Some counterarguments against prioritising risk modelling:
But overall, we think it is very worthwhile to work on Risk Modelling.
Obviously, creating more risk models. Currently, only a small portion of the threats posed by Frontier AI is covered by Risk Models. Many risks have been discussed for years, yet nobody has rigorously modeled them.
Some examples include:
A concrete way to do this is presented in Appendix 1.
Improving methodology. The current SOTA Risk Modelling methodology is nascent. Examples of methodological improvements are UK AISI’s structured protocol for Expert Elicitation, CARMA’s framework for Probabilistic Risk Assessment, or Safer AI’s research combining benchmark scores with expert elicitation. Further work is necessary to make risk modelling more scalable, navigate widely diverging expert opinions, appropriately include uncertainty, or employ principled methods of forecasting damages. (There are many other possible improvements.) It might be especially fruitful to study practices and frameworks from fields with more mature Risk Modeling, such as aviation or healthcare, including Fault Tree Analysis, Bow-Tie Analysis, or Key Risk Indicators.
Building consensus and coordination. If the scientific community can find agreement on which Risk Models and methodologies are SOTA, the AI Office could force companies to apply these. Such coordination could involve organizing workshops or conducting surveys of experts to determine which risk models and methodologies are regarded as the best. For example, helping the field converge on which evidence should be used in Risk Models (e.g., uplift studies, capability benchmarks, expert judgements).
If you are interested, you can complete this form to join a Slack channel to coordinate on this!
Mapping AI Benchmark Data to Quantitative Risk Estimates Through Expert Elicitation (Campos, 2025)
How do they go from a given level of capability to a level of risk:
Some orgs doing work on this include:
Some relevant papers:
Work done at the international SC-WOP workshop, 2025.