Estimating AI Harm Rates for Germany: Applying Epidemiological Methods to Incident Monitoring
Mentor: Branwen Nia Owen
Project area: The epidemiology of AI risk
Project Language
English or German.
Minimum Time Commitment
8 hours per week.
Project Abstract
AI incident databases are gaining traction, but raw counts don't tell policymakers much. A rise in reported harms could mean genuine risk escalation — or just improved reporting, or increased AI deployment. To distinguish these, you need both a numerator (how many harms actually occurred, adjusted for reporting gaps) and a denominator (how many opportunities for harm existed).
This project adapts epidemiological methods to estimate AI harm rates for incident types relevant to German policy priorities — such as hiring algorithms under anti-discrimination law, clinical decision support in the German healthcare system, or content moderation on platforms operating under the DSA.
The work has two components:
1. Exposure estimation: How many opportunities for harm existed? Depending on data availability, we use direct measurement (e.g., DSA transparency reports), category-based assignment, decomposition into bounded components, or structured expert judgment.
2. Harm estimation: How many incidents actually occurred? Reported counts underestimate true incidence. We apply methods from disease surveillance — such as capture-recapture across overlapping sources and multiplication factors from analogous domains — to adjust for underreporting.
Mentees will:
- Select incident types relevant to German AI governance;
- Estimate both exposure and harm counts, with documented assumptions;
- Calculate crude harm rates and assess trends;
- Produce standardised output cards suitable for policy communication;
Expected outputs include worked examples for German-relevant incident types, a practical methodology guide, and a gap analysis identifying where better data collection would most improve monitoring capacity.
Theory of Change
Effective AI governance requires prioritisation. Policymakers cannot address all AI harms simultaneously — they need to know which are escalating fastest and which are most severe relative to deployment scale.
Currently, AI incident databases provide raw counts, which conflate genuine risk changes with reporting improvements and deployment growth. This makes rational prioritisation difficult. Policymakers either react to media salience or treat all incident types as equally urgent.
This project provides the analytical layer that sits between raw incident data and policy decisions. By estimating harm rates (incidents per opportunity), we enable:
1. Trend detection: Distinguishing genuine escalation from increased reporting or deployment
2. Cross-type comparison: Identifying which incident types have the highest harm rates, not just the highest counts
3. Early warning: Detecting phase shifts before they become crises
For Germany specifically, this builds domestic capacity to interpret AI incident data rather than relying on US or EU-level analyses that may not reflect German deployment patterns or regulatory context.
The path to impact: methodology → worked examples → adoption by German policy analysts → better-informed prioritisation → more effective allocation of regulatory attention and intervention resources.
Desired Mentee Background
Any or all, it's more about skills and resourcefulness than a field of study.
Desired Mentee Level of Education
Masters and above.
Other Mentee Requirements
* Comfort with quantitative reasoning (e.g., interpreting rates, proportions, uncertainty)
* Familiarity with basic statistics (confidence intervals, bias, sampling, regression at a conceptual level)
* Ability to read technical papers and policy documents
* Comfort working under ambiguity and with incomplete data
* Ability to produce clear, policy-facing summaries (not just technical notes)
* Pragmatic, impact-oriented mindset