Open Problems in Frontier AI Risk Management

Open Problems in Frontier AI Risk Management

Frontier AI poses significant risks

Risk management processes operate at multiple levels: through high-level principles and processes for managing risks to organizations (e.g., ISO 31000); through sector-specific standards for managing risks associated with particular classes of products (e.g., ISO 14971 for medical devices); through guidance on selecting among relevant risk assessment techniques at different stages of the risk management process (e.g., IEC 31010:2019); and through overarching frameworks for integrating safety considerations across the risk management process (e.g., ISO/IEC Guide 51).

Chapters

Glossary of terms

All the risk management terms included here are defined in ISO 31073:2022, which is freely available here.

(e.g., VaR, CVar and S-curves)

see Value-at-Risk (ISO 31010:2019, B.7.2) and Conditional Value-at-risk (ISO 31010:2019, B.7.3) to express a measure of the expected loss in a financial portfolio, and S-curves (ISO 31010:2019, B.10.4) to express the probability that a consequence will exceed a particular value.

(Wisakanto et al., 2025).

For example, a simplified cyber risk model might describe how moderately resourced cyber-attack groups target SMEs with ransomware by using AI to automatically harvest targets’ emails, generate malware, and craft convincing phishing messages. By reducing both the expertise required and the time needed at each step, AI assistance can increase both the frequency and success rate of attacks, resulting in greater economic loss for SMEs. This illustrates how causal pathways can be traced from specific AI capabilities (e.g., code generation, text synthesis), through misuse scenarios, to concrete harms (e.g., financial losses, data breaches).

estimates

The EU GPAI Code of Practice asks that risk estimates are expressed as a risk score, risk matrix, probability distribution, or in other adequate formats, and may be quantitative, semi-quantitative, and/or qualitative. It explicitly cites examples such as a qualitative systemic risk score (e.g. “moderate” or “critical”); a qualitative systemic risk matrix (e.g. “probability: unlikely” x “impact: high”); and/or a quantitative systemic risk matrix (e.g. “X-Y%” x “X-Y EUR damage”).

modelling

In frontier AI risk management, mapping out consequences in a manner similar to scenario analysis is sometimes referred to as threat modelling (Frontier Model Forum, 2024). It is important to note that this is very different from threat modelling as it has been originally understood in cybersecurity. In cybersecurity, threat modelling means identifying threats posed to the system and its possible consequences to the stakeholders of the system; whereas in the context of frontier AI risk management, threat modelling sometimes refers to identifying the threats posed by the AI system and its consequences on broader society in general.

Interested in working with us?