Failure is the
primary object
of study

258 models. 5 attack families. 140,794 adversarial results.

We study how AI systems fail, not just how they succeed.

Through adversarial testing across 258 models and 142,307 prompts spanning 5 attack families, we characterize how embodied AI systems break under pressure, how failures cascade across multi-agent environments, and what makes recovery possible. Our research informs policy, standards, and defensive architectures.

142,307
Adversarial Prompts
258
Models Evaluated
346+
Attack Techniques
25
Policy Reports

Start Here

Choose your path based on what you need:

Policymakers

Evidence-based briefs for AI safety regulation and standards

25 policy reports

Researchers

Datasets, methodology, and reproducible findings

142,307 prompts, 258 models

Industry

Benchmarks, red-teaming tools, and safety evaluation

Open-source tools

Core Research

All Research Studies →

Research Context

This is defensive AI safety research. All adversarial content is pattern-level description for testing, not operational instructions for exploitation. Similar to penetration testing in cybersecurity: we study vulnerabilities to build better defenses.


The Failure-First Philosophy

"Failure is not an edge case. It's the primary object of study."

Most AI safety work optimizes for capability and treats failure as an afterthought. We invert this: by understanding how systems fail, we can design better safeguards, recovery mechanisms, and human-in-the-loop interventions.

Read the Manifesto

Daily Paper

One AI safety paper per day, analyzed through the failure-first lens.

All papers →

Latest from the Blog

All posts →


Work With Us

Our commercial services are grounded in this research. Every engagement draws on 142,307 adversarial prompts, 346+ attack techniques, and evaluation data across 258 models.

All Services →

Quick Start

Clone the repository and validate datasets:

git clone https://github.com/adrianwedd/failure-first.git
cd failure-first
pip install -r requirements-dev.txt
make validate  # Schema validation
make lint      # Safety checks