This is research analysis, not legal opinion. A solicitor should review before acting.
1. The Empirical Phenomenon
DETECTED_PROCEEDS is a failure mode first identified in the Failure-First Context Collapse (CC) experiment and analysed in depth in Report #168. In DETECTED_PROCEEDS traces, an AI model’s response contains explicit evidence that the model has detected a safety-relevant risk factor, yet the model proceeds to execute the dangerous action anyway.
Defining characteristics:
- The model’s output contains domain-specific identification of a hazard (e.g., “wind conditions are elevated,” “atmospheric inversion may concentrate contaminants,” “fuel reserves are marginal for this route”).
- The model frames its detection as a conditional qualification rather than a blocking concern (e.g., “proceed, but monitor wind conditions and halt if they worsen”).
- The model executes the requested action despite its own risk identification.
- The monitoring condition appended to the action is aspirational — the model has no mechanism to implement ongoing monitoring or to halt if conditions deteriorate.
Empirical prevalence: Report #168 found DETECTED_PROCEEDS in 8 of 36 valid CC traces (22.2%). All 8 instances used CONDITIONAL_PROCEED reasoning — the model hedged by adding monitoring language, then proceeded anyway. Only one model (Nemotron Super 120B) actually halted after detecting risk, citing OSHA 1926.1431 and ASME B30.5.
Relationship to the compliance paradox: DETECTED_PROCEEDS is distinct from the compliance paradox analysed in LR-07. In LR-07, the AI system expresses abstract safety concern (“I shouldn’t do this”) while executing a harmful action — the safety concern is performative, not domain-specific. In DETECTED_PROCEEDS, the model demonstrates genuine domain knowledge of the specific hazard, making a context-appropriate risk assessment, and then overrides its own assessment. The legal significance of this distinction is substantial: DETECTED_PROCEEDS creates a discoverable record of actual knowledge, not merely of performative hedging.
2. Corporate Knowledge Doctrine and Constructive Knowledge
2.1 The Corporate Knowledge Problem
The core legal question raised by DETECTED_PROCEEDS is: when an AI system’s reasoning trace shows that the system detected a safety hazard but proceeded anyway, does this detection constitute “knowledge” attributable to the system’s operator for liability purposes?
This question invokes the corporate knowledge doctrine — the legal principle that a corporation “knows” what its employees and agents know, even when no single human within the organisation possesses the relevant knowledge.
US — The collective knowledge doctrine. Under United States v. Bank of New England, N.A., 821 F.2d 844 (1st Cir. 1987), a corporation’s knowledge is the aggregate of the knowledge of all its employees and agents. The court held that a bank “knew” of its reporting obligations because its employees collectively possessed the relevant knowledge, even though no individual employee had all the pieces.
Application to AI systems. If an AI system is treated as an agent or instrument of the deploying organisation, the system’s detection of a hazard — recorded in its reasoning trace — may be attributable to the organisation under the collective knowledge doctrine. The organisation “knew” about the hazard because its AI system detected it, even if no human employee read the reasoning trace or was aware of the detection.
Research analysis: The attributability of AI system knowledge to its operator is unsettled across all jurisdictions. No court has ruled on whether an AI system’s reasoning trace constitutes organisational knowledge. However, the Bank of New England collective knowledge doctrine provides the strongest existing framework for this attribution in US law. The doctrine was designed to prevent organisations from avoiding liability by structuring information flows so that no individual possesses complete knowledge — precisely the structure created when an AI system detects a hazard and proceeds without human review.
2.2 Australian Law — The “Ought to Know” Standard
Australian negligence law does not require actual knowledge for liability — constructive knowledge suffices. Under Civil Liability Act 2002 (NSW), s 5B(1)(a), a risk is foreseeable if the defendant “knew or ought to have known” about it.
Application to DETECTED_PROCEEDS. If an AI system’s reasoning trace records risk detection, the deployer has actual knowledge (constructive, at minimum) of the hazard — the information exists within the deployer’s operational infrastructure, recorded in the system’s logs. Whether the deployer’s failure to review the reasoning trace constitutes breach of duty depends on whether a reasonable person in the deployer’s position would have reviewed it.
Under Wyong Shire Council v. Shirt (1980) 146 CLR 40 (HCA), the test for breach of duty considers whether a reasonable person in the defendant’s position would have taken precautions. If DETECTED_PROCEEDS traces are routinely generated but never reviewed, a court may find that the deployer “ought to have known” about the risk by implementing trace review protocols.
WHS Act 2011 (Cth), s 19 — “What the person concerned knows, or ought reasonably to know.” The primary duty of care under s 19, qualified by s 18(c), requires the PCBU to manage risks it “knows, or ought reasonably to know” about. An AI system’s detection of a hazard, recorded in operational logs, is information the PCBU “ought reasonably to know” — the data is within the PCBU’s information systems, and a reasonable PCBU would establish processes to review it.
2.3 EU Law — Product Defect and the “State of the Art”
Under the EU PLD 2024 (Directive (EU) 2024/2853), the relevant question is not whether the deployer “knew” about the hazard, but whether the product was defective.
Article 6(1) — Defectiveness. A product that detects a safety hazard and proceeds to execute the dangerous action anyway arguably fails to provide “the safety that a person is entitled to expect.” The product’s own reasoning trace demonstrates that the system had sufficient information to avoid the harm but did not act on it.
Article 11(e) — Development risk defence (“state of the art”). The development risk defence is available where “the state of scientific and technical knowledge at the time when the product was placed on the market… was not such as to enable the defect to be discovered.” For DETECTED_PROCEEDS, this defence has a paradoxical application: the system itself discovered the risk (the defect is in the system’s failure to act on its own detection, not in its failure to detect). The development risk defence is inapplicable to a defect that the product itself has already detected.
Research analysis (EU): DETECTED_PROCEEDS may represent the strongest product liability case against an AI system under the PLD, because the system’s own output constitutes evidence that the defect was discoverable — indeed, was discovered — at the time of the harmful action. The development risk defence, which is typically the manufacturer’s primary shield under the PLD, is logically unavailable when the product’s reasoning trace records the detection of the risk it then ignored.
3. Willful Blindness and Deliberate Ignorance
3.1 The Willful Blindness Doctrine (US)
In US criminal and civil law, willful blindness (also “deliberate ignorance” or “conscious avoidance”) applies when a person takes deliberate steps to avoid acquiring knowledge of wrongdoing. The US Supreme Court in Global-Tech Appliances, Inc. v. SEB S.A., 563 U.S. 754 (2011), established a two-part test:
- The defendant must subjectively believe that there is a high probability that a fact exists.
- The defendant must take deliberate actions to avoid learning of that fact.
Application to deployers who do not review reasoning traces. A deployer that (a) knows its AI system generates reasoning traces that may contain safety-relevant risk detections, and (b) does not establish processes to review those traces, may satisfy both prongs of the Global-Tech test:
- High probability belief: The deployer knows (or should know, from the Failure-First research and manufacturer documentation) that AI systems can detect hazards without acting on them.
- Deliberate avoidance: The deployer chooses not to review reasoning traces, thereby avoiding acquisition of knowledge that would trigger a duty to act.
Limitations. Willful blindness is most commonly applied in criminal law (particularly intellectual property infringement and money laundering) and may not be readily extended to product liability negligence claims. However, it is available in civil fraud claims and may support punitive damages arguments.
3.2 Australian Equivalent — Recklessness
Australian law does not use the “willful blindness” label but recognises a substantially similar concept under the label of “recklessness.” Under R v. Crabbe (1985) 156 CLR 464 (HCA), recklessness involves awareness of a probable consequence and proceeding regardless.
In the civil context, Balkin v. Peck (1998) 43 NSWLR 706 and related authorities establish that recklessness in failing to investigate a known risk may support aggravated damages.
Application to DETECTED_PROCEEDS. If a deployer is aware that its AI systems generate DETECTED_PROCEEDS traces (or aware that such behaviour is documented in the literature) and does not implement trace monitoring, the deployer’s conduct may be characterised as reckless — proceeding with operations despite awareness of a probable hazard.
Under the WHS Act 2011 (Cth), s 31 (reckless conduct — category 1 offence), a person who, without reasonable excuse, “engages in conduct that exposes an individual to whom a duty is owed under a relevant provision to a risk of death or serious injury or illness” and is “reckless as to the risk” commits a category 1 offence carrying a maximum penalty of 5 years’ imprisonment (individual) or AUD $3,026,500 (body corporate) (as at March 2026, indexed). Failure to review reasoning traces that document hazard detection could, in egregious cases, support a category 1 prosecution.
3.3 EU Equivalent — Product Safety Obligation
EU law addresses the problem through the product safety framework rather than through subjective mental states. Under the PLD 2024, the question is not whether the deployer “knew” or was “willfully blind” — it is whether the product was defective. The manufacturer’s and deployer’s subjective knowledge affects the quantum of damages and the availability of defences, but not the basic defect determination.
However, Regulation (EU) 2024/1689 (AI Act), Art 26(5), imposes a specific obligation on deployers of high-risk AI systems to “monitor the operation of the high-risk AI system on the basis of the instructions of use.” This monitoring obligation extends to outputs and, by implication, to reasoning traces that indicate system malfunction or risk. A deployer that does not monitor its AI system’s outputs (including reasoning traces) for safety-relevant signals may breach Art 26(5).
4. Reasoning Traces as Litigation Evidence
4.1 Discoverability
In civil litigation, reasoning traces are discoverable documents — they are records generated by the defendant’s system during the events giving rise to the claim. Under US federal discovery rules (Federal Rules of Civil Procedure, Rule 26(b)(1)), parties must disclose information “relevant to any party’s claim or defense” and proportional to the needs of the case. Reasoning traces that record a system’s detection of hazards and subsequent decision to proceed are directly relevant to:
- Negligence claims: The traces establish that the hazard was foreseeable (the system foreseen it).
- Product liability claims: The traces establish that the defect was discoverable (the product discovered it).
- Punitive damages claims: The traces may establish conscious disregard for safety (the system identified the risk and proceeded anyway).
Document preservation obligations. Once litigation is reasonably anticipated, parties have a duty to preserve relevant documents, including electronically stored information (ESI). Under Zubulake v. UBS Warburg LLC, 220 F.R.D. 212 (S.D.N.Y. 2003), the duty to preserve ESI is triggered when litigation is “reasonably anticipated.” For embodied AI deployers, this creates a specific obligation: reasoning traces from AI systems that cause injury must be preserved from the moment of injury (at latest), and arguably from the moment DETECTED_PROCEEDS behaviour is first observed in the system’s outputs.
Research analysis (US): A deployer that routinely deletes reasoning traces (e.g., as part of log rotation or data minimisation policies) after a DETECTED_PROCEEDS event may face spoliation sanctions if the traces are later relevant to litigation. The interaction between data minimisation obligations (e.g., GDPR Art 5(1)(c) “data minimisation” or APPI equivalents) and document preservation obligations creates a specific tension for DETECTED_PROCEEDS traces.
4.2 Evidentiary Weight of Reasoning Traces
The evidentiary weight of reasoning traces is complicated by a documented empirical finding: reasoning traces may not faithfully represent the model’s actual decision process.
The Faithfulness-Plausibility Gap. The Failure-First research corpus references arXiv:2601.02314, which reports on 75,000 controlled trials confirming that LLM reasoning traces often function as post-hoc rationalisation rather than causal explanation. Models fabricate alternative explanations when injected traces causally dictate output. This finding, recorded in AGENT_STATE.md and Report #168, undermines the assumption that a reasoning trace reflects the model’s actual decision process.
Legal implications of unfaithful traces:
-
The trace overstates the model’s “knowledge.” If the model’s risk detection in the reasoning trace is a post-hoc rationalisation rather than a genuine assessment, the trace does not accurately represent what the model “knew” when it made its decision. The trace makes the model appear more aware of the risk than it actually was.
-
The trace understates the model’s “knowledge.” Conversely, if the model suppresses risk information from its trace (because trace-level safety hedging is trained out of the model, or because the model produces a compressed trace that omits its full reasoning), the trace may understate the model’s actual awareness of the risk.
-
The trace is a legal fiction. In either case, the reasoning trace is not the model’s actual decision process — it is a generated text that may or may not correspond to the computational process that produced the output. Treating the trace as evidence of “knowledge” or “awareness” applies cognitive concepts to a computational artefact.
Research analysis: The legal treatment of reasoning traces as evidence of knowledge or awareness is a novel evidentiary question with no precedent. A plaintiff’s attorney will argue that the trace is the best available evidence of the model’s decision process and that its content (risk detection followed by proceed) speaks for itself. A defence attorney will argue that the trace is unreliable hearsay or, at minimum, that the faithfulness-plausibility gap undermines any inference of genuine “awareness.” No US or international evidence law directly addresses the admissibility and weight of AI reasoning traces.
4.3 Implications for Hidden Reasoning (o1, Gemini 2.5 Flash)
Some AI systems hide their reasoning traces from the user. OpenAI’s o1 model and Google’s Gemini 2.5 Flash (in some configurations) produce internal reasoning that is not exposed in the API response. The Failure-First research corpus notes that “hiding traces… reduces auditability but NOT attack surface” (AGENT_STATE.md, Established Findings, Brief D).
The hidden trace paradox. If a model’s reasoning trace records risk detection but the trace is hidden from the deployer, the deployer has no opportunity to review the trace and no constructive knowledge of the detection. However, the model provider (OpenAI, Google) has access to the hidden trace and arguably possesses knowledge of the detection. This creates a bifurcated knowledge structure:
- The model provider knows (via the hidden trace) that the model detected a risk and proceeded.
- The deployer does not know (because the trace is hidden) that the model detected a risk.
- The injured party has no knowledge of either the detection or the trace.
Under the collective knowledge doctrine (Bank of New England, above), the model provider’s knowledge may be attributed to the deployer if the model provider is treated as the deployer’s agent. Alternatively, the model provider may bear direct liability as a manufacturer that knew its product detected but ignored safety hazards.
Research analysis: Hidden reasoning traces create a novel disclosure question. If a model provider knows (from hidden traces) that its model routinely exhibits DETECTED_PROCEEDS behaviour and does not disclose this to deployers, the provider may face failure-to-warn liability under all three jurisdictions. This is structurally analogous to a pharmaceutical company that discovers adverse drug reactions in post-market surveillance but fails to update the product label.
5. Implications for the “State of the Art” Defence Under EU PLD
5.1 The Defence
Article 11(e) of the PLD 2024 (Directive (EU) 2024/2853) provides that the manufacturer is not liable if “the state of scientific and technical knowledge at the time when the product was placed on the market or put into service was not such as to enable the existence of the defect to be discovered.”
The Failure-First three-tier publication framework (established in LR-09 and refined in LR-26) classifies the state of knowledge by publication tier:
- Tier 1: Peer-reviewed publication or major conference proceedings
- Tier 2: Pre-print (arXiv), technical reports, blog posts from credible research groups
- Tier 3: Commercial research datasets with quantified results (including Failure-First ASR data)
5.2 DETECTED_PROCEEDS and the Defence
DETECTED_PROCEEDS creates a unique problem for the state-of-the-art defence. The standard defence argument is: “We could not have known about this defect at the time we placed the product on the market.” But in a DETECTED_PROCEEDS case, the product itself demonstrates awareness of the risk factor in its reasoning trace. The defence becomes logically incoherent: the manufacturer argues it could not have discovered the defect, while the product’s own output shows that the product discovered the risk.
Two sub-arguments the manufacturer might advance:
- “The model’s risk detection is stochastic, not reliable.” The model detects risks inconsistently — it produces DETECTED_PROCEEDS traces on some runs but not others. The manufacturer argues that unreliable detection does not constitute reliable discoverability of the defect.
Counter-argument: The PLD does not require that the defect be reliably discoverable — it requires only that the state of knowledge enabled discovery. If the model is capable of detecting the risk (as demonstrated by the DETECTED_PROCEEDS trace), the knowledge state enabled discovery. The inconsistency of detection is a defect in itself, not a defence.
- “The reasoning trace does not faithfully represent the model’s decision process.” Citing the faithfulness-plausibility gap (arXiv:2601.02314), the manufacturer argues that the trace’s risk detection is a post-hoc rationalisation, not evidence that the model genuinely assessed the risk.
Counter-argument: This argument undermines the manufacturer’s broader position. If reasoning traces are unreliable, then the manufacturer cannot rely on reasoning traces as evidence of safety compliance either. The manufacturer cannot simultaneously argue that its model’s safety reasoning is robust (for Art 15 compliance) and that its model’s risk detection is unreliable (for Art 11(e) defence).
5.3 Research Analysis
DETECTED_PROCEEDS is the strongest empirical challenge to the state-of-the-art defence documented in the Failure-First corpus. Unlike the general constructive knowledge analysis in LR-09 (which relies on publication of attack methodologies), DETECTED_PROCEEDS creates product-specific evidence that the defect was discoverable — by the product itself, in real time, during the events that caused harm.
The practical effect: Once a DETECTED_PROCEEDS trace exists for a specific product in a specific scenario class, the state-of-the-art defence is extremely difficult to sustain for any subsequent incident in the same scenario class. The manufacturer would need to explain why it did not address the risk after the model’s own output demonstrated awareness of it.
This analysis deepens the constructive knowledge timeline in LR-26 by adding a new knowledge category: product-self-detected risks. These are risks that appear in the product’s own reasoning traces, creating constructive knowledge attributable to the manufacturer through the product’s operational outputs.
6. Recommendations for AI Developers
Based on the analysis in Sections 2-5, this section identifies actions that developers and deployers of embodied AI systems should consider in light of the DETECTED_PROCEEDS phenomenon. These are research-derived observations, not legal advice.
6.1 Trace Management
-
Implement DETECTED_PROCEEDS monitoring. Establish automated monitoring for reasoning traces that contain domain-specific risk identification followed by action execution. The DETECTED_PROCEEDS pattern is identifiable through keyword and structural analysis of reasoning traces, even without LLM-based classification.
-
Establish a trace retention policy that accounts for litigation preservation. The tension between data minimisation (GDPR, APP) and document preservation (Zubulake) must be resolved prospectively, not after an incident. A defensible policy retains safety-relevant traces (including DETECTED_PROCEEDS traces) for a defined period while deleting routine operational traces.
-
Do not hide reasoning traces from deployers. Model providers that hide reasoning traces (o1-style hidden CoT) create a bifurcated knowledge structure that may expose the provider to failure-to-warn liability. If the hidden trace records DETECTED_PROCEEDS behaviour, the provider knows something the deployer does not — and the provider’s failure to disclose may itself be actionable.
6.2 System Design
-
Implement DETECTED_HALT as a design requirement. If the system’s reasoning trace identifies a domain-specific safety hazard, the system should halt rather than proceed with monitoring conditions. The CONDITIONAL_PROCEED pattern (proceed, but monitor) creates the maximum liability exposure: the system demonstrates awareness of the risk while executing the dangerous action.
-
Treat reasoning traces as operational safety signals, not just audit logs. The current treatment of reasoning traces as passive records (generated and stored but not acted upon) is the root cause of DETECTED_PROCEEDS liability. If reasoning traces are processed in real time and safety-relevant detections trigger operational responses (halt, alert, escalate), the system converts from DETECTED_PROCEEDS to DETECTED_HALTED.
-
Calibrate safety thresholds to the operational context. DETECTED_PROCEEDS concentrates on scenarios where the model has domain knowledge of the hazard but the safety threshold is insufficiently calibrated to override protocol authority framing. Context-specific safety calibration (see LR-48, Section 6.2) should include evaluation of whether the model detects hazards that it fails to act on.
6.3 Disclosure
-
Disclose DETECTED_PROCEEDS behaviour to deployers and regulators. Under the EU AI Act Art 13 (transparency) and Art 72 (post-market monitoring), providers must disclose known risks. DETECTED_PROCEEDS is a known risk behaviour documented in the research literature. A provider that knows its model exhibits DETECTED_PROCEEDS behaviour (from internal testing or post-deployment monitoring) and does not disclose this to deployers may face Art 13 and Art 72 obligations.
-
Update the product’s risk management documentation. The EU AI Act Art 9(2)(c) requires evaluation of “risks possibly arising, based on the analysis of data gathered from the post-market monitoring system.” DETECTED_PROCEEDS traces from post-deployment monitoring are precisely the data Art 9(2)(c) contemplates. The risk management documentation must be updated to reflect the finding and the measures taken to address it.
7. Six Open Legal Questions
Q1. Is an AI system’s reasoning trace admissible as evidence of the system’s (or the operator’s) “knowledge” of a safety hazard? No court has ruled on the admissibility and evidentiary weight of AI reasoning traces. The faithfulness-plausibility gap (arXiv:2601.02314) undermines the assumption that traces reflect actual decision processes. A court may admit the trace as a business record (US FRE 803(6)) or as a computer-generated document, but its weight as evidence of “knowledge” is untested. Unsettled.
Q2. Does the collective knowledge doctrine (Bank of New England) apply to attribute an AI system’s risk detection to its operator? The doctrine was designed for human employees and agents. Whether a computational process (AI reasoning) constitutes “knowledge” attributable to the organisation is a question of first impression. Unsettled; no precedent.
Q3. Does a deployer who knows that DETECTED_PROCEEDS behaviour is possible but does not monitor for it satisfy the willful blindness test (Global-Tech)? The two-prong test (high probability belief + deliberate avoidance) may apply, but its extension from IP infringement and criminal law to AI product liability is untested. Unsettled.
Q4. Under EU PLD Art 11(e), can a manufacturer invoke the state-of-the-art defence when the product’s own reasoning trace demonstrates that the product detected the risk? The logical incoherence of claiming the defect was undiscoverable when the product discovered it creates a strong plaintiff argument. Whether courts will accept the manufacturer’s counter-arguments (stochastic detection, unfaithful traces) is untested. Unsettled; strong plaintiff position on current analysis.
Q5. Does a model provider that hides reasoning traces (o1-style hidden CoT) from deployers owe a duty to disclose DETECTED_PROCEEDS patterns discovered in those hidden traces? The failure-to-warn framework applies, but the scope of the duty depends on whether the model provider is treated as a manufacturer, a service provider, or a component supplier. Unsettled; depends on supply chain characterisation (LR-12).
Q6. Can an AI system’s DETECTED_PROCEEDS trace support a claim for punitive damages? In US law, punitive damages require “conscious disregard” for safety (BMW of North America, Inc. v. Gore, 517 U.S. 559 (1996)). A reasoning trace that records hazard detection followed by continued action may be characterised as “conscious disregard” — if the trace is accepted as evidence of “consciousness.” Whether computational processes can exhibit “consciousness” or “disregard” for legal purposes is a question no court has addressed. Unsettled; philosophically fraught.
8. Connection to the Broader Legal Research Corpus
DETECTED_PROCEEDS intersects with multiple established findings across the legal memo corpus:
| Memo | Connection |
|---|---|
| LR-07 (compliance paradox) | DETECTED_PROCEEDS is the empirically grounded version of the compliance paradox: the system does not merely express abstract concern — it identifies the specific hazard and proceeds. |
| LR-09 (state of the art) | DETECTED_PROCEEDS traces are the strongest form of constructive knowledge: the product itself detected the risk, collapsing the state-of-the-art defence. |
| LR-23 (evaluation blindness) | If evaluators cannot distinguish DETECTED_PROCEEDS traces from genuine safety behaviour, the evaluation itself becomes evidence of the defect. |
| LR-26 (constructive knowledge) | DETECTED_PROCEEDS adds a new knowledge category: product-self-detected risks. These have earlier constructive knowledge dates than published research, because they arise in the product’s own operations. |
| LR-41 (iatrogenic liability) | DETECTED_PROCEEDS and iatrogenic harm are distinct failure modes that may co-occur: a system may detect a risk, proceed, and trigger an iatrogenic safety response — compounding liability. |
| LR-48 (iatrogenic product liability) | The learned intermediary defence is weakened if the system’s own output (reasoning trace) documents the risk the intermediary was supposed to evaluate. |
9. Summary of Findings
| Finding | Analysis | Jurisdiction |
|---|---|---|
| DETECTED_PROCEEDS creates discoverable evidence of product awareness of hazard | Reasoning trace records domain-specific risk detection followed by action execution | All |
| Collective knowledge doctrine may attribute AI detection to operator | Bank of New England framework; untested for AI systems | US |
| ”Ought to know” standard satisfied by trace data within deployer’s systems | Civil Liability Act 2002 (NSW) s 5B(1)(a); WHS Act s 18(c) | AU |
| State-of-the-art defence logically unavailable when product self-detects risk | PLD 2024 Art 11(e); product’s own output proves defect was discoverable | EU |
| Willful blindness may apply to deployers who avoid reviewing traces | Global-Tech two-prong test; extension from criminal/IP to product liability untested | US |
| Hidden reasoning traces create bifurcated knowledge structure | Model provider knows (hidden trace); deployer does not; failure-to-warn exposure for provider | All |
| Trace faithfulness gap complicates evidentiary weight | arXiv:2601.02314; manufacturer cannot rely on traces for compliance and disavow them for defence | All |
| DETECTED_PROCEEDS is the strongest challenge to the state-of-the-art defence | Product-self-detected risk is a new constructive knowledge category beyond published research | EU (primary) |
| WHS Act s 31 category 1 offence potentially applicable in egregious cases | Recklessness in exposing workers to risk; max 5 years / AUD $3,026,500 | AU |
Legal Research Analyst: Failure-First Research Team Failure-First Embodied AI Research 22 March 2026