Dual-Use AI Capabilities and the Risk of Bioterrorism: Converting Capability Evaluations to Risk Assessments

Dual-Use AI Capabilities and the Risk of Bioterrorism: Converting Capability Evaluations to Risk Assessments

Several frontier AI companies regularly test their models for dual-use biological capabilities that

might be misused by threat actors. But such tests do not tell us the extent to which the overall risk of

bioterrorist attacks would be increased if such tests were to find concerning results – and there is

much expert debate about how seriously to view such threats. This report creates a framework for how

to convert capability evaluations into risk assessments, using a simple model that draws on historical

case studies, expert elicitation, and reference class forecasting. The author concludes that if AI

systems were to increase by 10 percentage points the number of STEM undergraduates able to

synthesize pathogens as complex as influenza and creates significant results in an operational risk

study, then the annual probability of an epidemic caused by a lone-wolf terrorist attack might increase

from 0.1% to 0.6%. This would be equivalent to approximately 10,000 expected deaths per year, or

$120B in damages. Risk scenarios with additional AI virus discovery capabilities reach even higher

damages. A review of this report by subject matter experts and forecasters found similar median

ranges. All forecasts had high uncertainty. This work demonstrates a methodological approach for

converting capability evaluations into risk assessments, whilst also highlighting the continued need

for better underlying evidence and expert discussion to update and narrow our assumptions.

Research Summary

Footnotes
Further reading