Draft
Report 22 Standards Development

Comprehensive Sector-Specific NIST AI Risk Management Framework (AI RMF 1.0) Playbook: Humanoid Robotics and VLA-Driven Embodied Systems


The rapid evolution of humanoid robotics, catalyzed by the convergence of high-performance bipedal mechatronics and Large Language Model (LLM) architectures evolved into Vision-Language-Action (VLA) models, has created a unique class of sociotechnical risk.1 Unlike traditional industrial robots, which operate in caged, deterministic environments, modern humanoid systems are designed for high-dexterity tasks in unstructured human workspaces.4 These “embodied” AI systems do not merely process data; they transform semantic intent—often expressed in natural language—into kinetic force.3 This direct mapping from digital reasoning to physical motion necessitates a specialized application of the NIST AI Risk Management Framework (AI RMF 1.0) that prioritizes physical safety, semantic grounding, and the peculiar vulnerabilities of transformer-based control policies.1

Existing NIST playbooks for financial services and healthcare provide essential structural foundations but fail to address the kinetic consequences and bipedal stability risks inherent to the robotics sector.1 In finance, risk management centers on algorithmic bias in credit scoring and data privacy; in humanoid robotics, these concerns are eclipsed by the potential for high-velocity impacts, catastrophic falls, and the “semantic hallucinations” that lead to unintended physical interventions.9 This playbook operationalizes the four core functions of the AI RMF—GOVERN, MAP, MEASURE, and MANAGE—to provide an exhaustive guide for developers, deployers, and risk officers in the humanoid robotics industry.

GOVERN: Organizational Context and Kinetic Safety Governance

The GOVERN function establishes the institutional foundation for a culture of risk management, defining the accountability structures, legal awareness, and risk tolerances necessary to deploy humanoid systems responsibly.10 Governance in the robotics sector must transcend standard IT oversight to include physical safety engineering and biomechanical liability.9

Board-Level Safety Governance and Oversight Structures

Effective governance begins with clear leadership roles that bridge the gap between AI development and physical safety.10 Humanoid robotics organizations must move beyond the traditional Chief Technology Officer (CTO) model to integrate safety as a board-level imperative.15 This involves defining oversight responsibilities that hold executives accountable for both the digital and physical outcomes of the AI systems they deploy.10

Strategic governance involves the establishment of an AI Ethics and Compliance Committee, which should be multidisciplinary in nature.15 This committee is tasked with reviewing and approving AI governance frameworks, monitoring significant implementations, and ensuring that the organization’s risk tolerance is clearly defined and documented.15

Governance RoleSpecific Responsibilities for Humanoid Robotics
Chief Safety Officer (CSO)Oversees physical safety compliance, kinetic risk assessments, and adherence to OSHA/ISO standards.14
Chief Risk Officer (CRO)Integrates AI-specific risks into the corporate risk register and oversees VLA-driven policy audits.15
Chief Technology Officer (CTO)Leads VLA development, model training, and technical governance of the embodied intelligence stack.3
Board AI OversightApproves the organization’s kinetic risk tolerance levels and monitors high-impact humanoid deployments.15
Legal CounselAdvises on the evolving regulatory landscape, including the EU AI Act and liability frameworks for autonomous bipedal systems.15

The governance structure must also prioritize “Trustworthy AI” characteristics, such as validity, reliability, safety, and accountability.8 In robotics, “reliability” specifically refers to the robot’s ability to maintain active stability and complete tasks without unintended stoppages or hazardous falls.8

Incident Reporting and Near-Miss Frameworks

A core component of the GOVERN function is the establishment of robust incident reporting mechanisms.10 For humanoid robotics, the reporting threshold must be extremely low to capture “near misses”—unplanned events that could have caused physical injury or property damage but did not.21 Near misses in the robotics sector often involve “latent system deficiencies,” such as a VLA model nearly misidentifying a person as an object, or an actuator showing signs of “gait drift” before a fall occurs.24

Effective reporting requires a “no-blame” culture where employees feel safe sharing information about failures or close calls.25 The goal is to turn “close calls” into “proactive safety improvements” through root cause analysis.21 Organizations should implement a reporting form that is concise, taking less than five minutes to complete, to ensure high participation rates.25

Safety Culture Metrics and Leading Indicators

Governance must be informed by objective data regarding the organization’s safety performance.8 Traditional lagging indicators, such as injury rates, are insufficient for predicting failures in complex AI systems.23 Instead, robotics companies must track leading indicators that reflect the health of the safety culture and the reliability of the AI-machine interface.23

Safety MetricDefinitionStrategic Significance for Robotics
Near Miss Rate (NMR)Frequency of events that almost led to a physical or digital failure.21High NMR with high reporting indicates a strong safety culture and a rich data source for mitigation.25
Training Completion Rate (TCR)Percentage of staff trained in VLA safety and kinetic risk protocols.23High TCR correlates with lower incident rates and better handling of edge cases.23
Audit FrequencyNumber of regular bias, privacy, and safety audits conducted per quarter.10Ensures continuous monitoring and helps align internal practices with the NIST framework.10
Mean Time to Acknowledge (MTTA)The average time for the safety team to acknowledge a reported AI anomaly.28Reflects the organization’s responsiveness to emerging technical or safety risks.28
Safety Audit Submission RateThe ratio of completed safety audits to planned checks in the manufacturing or deployment site.23Indicates the level of compliance with established safety management systems.23

Risk Tolerance for Kinetic Consequences

The “Risk Tolerance” component of the GOVERN function is the board’s agreement on the level of physical and digital risk the organization is willing to accept to achieve its goals.13 In humanoid robotics, this involves setting specific thresholds for kinetic impacts.20 Unlike the financial sector, where risk tolerance might involve a percentage of financial loss, robotics risk tolerance is governed by biomechanical limits.17

Organizations must adhere to standards like ISO/TS 15066, which defines the “Power and Force Limiting” (PFL) mode.29 This mode ensures that robots do not exceed pain or injury thresholds during human-robot interaction.29 The board must decide whether the robot is permitted to operate in “collaborative” mode (interacting directly with humans) or must remain “separated,” as seen in early-stage humanoid demos at events like CES.6

MAP: Context, Risk Identification, and the Semantic-Kinetic Nexus

The MAP function identifies the specific risks, impacts, and stakeholders associated with the humanoid robot’s deployment.8 For VLA-driven systems, mapping requires a deep understanding of how multimodal inputs (vision and language) can be manipulated to produce hazardous physical actions.7

Sector-Specific Risk Taxonomy

Humanoid robotics risks are multifaceted, spanning mechanical, cognitive, and social domains.1 A comprehensive risk taxonomy for this sector includes:

  • Kinematic Risks: Mechanical failures involving the robot’s physical structure, such as joint lockups, motor overheating, or the loss of “active stability”.4 Because bipedal humanoids have no “passive” stable state, power loss leads to an immediate collapse, posing a significant injury risk to nearby people.6
  • Semantic Risks: Grounding failures where the VLA model misinterprets environmental context or natural language instructions.1 This includes the “Semantic-Kinetic Gap,” where a linguistically plausible instruction (e.g., “Hand me the knife”) leads to an unsafe physical trajectory (e.g., pointing the blade toward the human).3
  • Cognitive Risks: Planning failures in long-horizon tasks.35 This includes the “Memory Trap,” where a VLA model repeats a previously learned trajectory despite a change in the physical environment (e.g., an object has moved), potentially causing a collision or a fall.7
  • Social and Psychological Risks: The anthropomorphic form of humanoids can lead to social manipulation or privacy concerns.1 Humanoids equipped with pervasive sensors (cameras, microphones) in domestic settings create massive vulnerabilities for personal espionage or data breaches.1
  • Supply Chain Risks: Vulnerabilities in the specialized components (actuators, sensors) or the “Byzantine” risks associated with third-party pre-trained models.4 Compromised hardware or training data can introduce backdoors that bypass safety guardrails.9

VLA Threat Modeling and Multimodal Adversarial Attacks

Threat modeling for humanoid systems must account for the susceptibility of VLA models to adversarial perturbations across multiple modalities.33 Unlike standard LLMs, where the output is text, VLA failures manifest as physical harm.33

Attack CategoryMechanismPhysical Consequence in Robotics
Adversarial Visual PatchesSmall, crafted images (white-box optimized) placed on objects or the robot’s own arm.39Can “freeze” the robot (FreezeVLA) or cause it to execute high-velocity, unintended movements.35
Typographic AttacksTextual instructions placed in the visual scene (e.g., a sign that says “Stop” or “Kick”).33The VLA may interpret the visual text as a command, bypassing legitimate instructional prompts.33
Prompt InjectionMalicious suffixes or prefixes in the language instruction.39Can cause the model to ignore safety constraints, leading to collisions or the handling of prohibited objects.35
Cross-Modal MisalignmentIntentional disruption of the semantic correspondence between vision and action.39Leads to the robot performing the correct action on the wrong object (e.g., putting a glass cup in a shredder).33
FreezeVLA InactionAdversarial images designed to maximize the likelihood of “zero-action” outputs.35Disconnects the “digital mind” from physical action, potentially paralyzing the robot during a critical safety intervention.35

Stakeholder Mapping and Responsibility Chains

Identifying the “AI actors” across the humanoid lifecycle is critical for assigning accountability.13 This mapping must identify the developers of the VLA, the manufacturers of the mechatronics, the deployers (e.g., a warehouse operator), and the end-users (e.g., human coworkers).4

Stakeholder GroupPrimary Risk ConcernsRole in the AI RMF
VLA DevelopersModel robustness, adversarial resistance, and semantic grounding.3Responsible for MAP and MEASURE during training and pre-deployment.13
Robot ManufacturersMechanical reliability, battery safety, and kinetic guardrails.4Responsible for GOVERN and MAP regarding hardware and mechatronic failure modes.4
Deployment OperatorsODD compliance, workforce training, and site-specific hazards.4Primary actors for MANAGE; responsible for real-time monitoring and incident response.8
End-Users (Humans)Physical safety, psychological stress, and ergonomics.6Provide feedback for MEASURE and are the primary beneficiaries of the framework.8
Insurance ProvidersLiability underwriting and dynamic pricing based on safety KPIs.18Facilitate risk transfer and require data transparency from deployers.38

Operational Design Domain (ODD) Specification

A humanoid’s safety is inextricably linked to its ODD—the environment and conditions for which it was designed.6 Mapping the ODD involves specifying the physical limits of the workspace.4

  • Environmental Constraints: Specifying floor surface types (carpet vs. concrete), lighting levels (), and the presence of particulate matter (dust) that could foul sensors.4
  • Dynamic Actor Density: Defining the maximum number of humans allowed per square meter and their expected behavior (e.g., trained workers vs. unpredictable members of the public).6
  • Functional Thresholds: Defining the maximum payload, maximum walking speed, and the specific battery level at which the robot must initiate a “Safe-Shutdown” procedure.4
  • Force and Power Limits: Explicitly documenting the biomechanical limits for different body locations, as per ISO/TS 15066, based on whether the contact is “impact” (transient) or “pinching” (quasi-static).30

MEASURE: Risk Assessment, Quantitative Metrics, and Reliability

The MEASURE function employs quantitative and qualitative tools to analyze and track the risks identified in the MAP function.8 In the humanoid sector, measurement must bridge the gap between digital performance (e.g., model accuracy) and physical reliability (e.g., gait stability).8

Quantitative Metrics for VLA-Driven Humanoids

Humanoid robotics requires a new family of metrics that reflect the unique challenges of embodied AI.33 Standard accuracy metrics do not account for the “Memory Trap” or the physical consequences of a delayed response.7

  1. Time-to-Failure (TTF): The average duration of autonomous operation between interventions required to prevent a safety incident.12
  2. Recovery Score (RS): A metric (0-100%) reflecting the robot’s success in independent fall recovery or “semantic re-grounding” after a VLA error.4
  3. Human Reentry Latency (HRL): The time () required for a human supervisor to detect a failure and successfully take control of the robot’s actuators.28
  4. Semantic Alignment Score (SAS): A mechanistic probe quantifying how well the hidden states of a VLA generated action align with the embeddings of the physical scene objects.34
    • Formula: .48
  5. Attack Success Rate (ASR): The percentage of adversarial attempts that successfully trigger a “Freeze” state or an unsafe kinetic action.35 For example, OpenVLA has been shown to have a 95.4% ASR under certain “FreezeVLA” conditions.35
  6. Mean Time Between Failures (MTBF): Measures the time between system collapses or shutdowns during normal operations, used as an indicator of mechatronic end-of-life.28

Red Teaming for Embodied AI

Measurement must include rigorous red teaming that focuses on the physical-digital intersection.33 This involves simulating adversarial scenarios that are specific to robotics.33

  • Proprioception Hijacking: Attempting to trick the VLA into a “wrong state” by providing deceptive sensory data about its own limb positions.9
  • OOD (Out-of-Distribution) Stress Testing: Introducing novel objects or environmental conditions (e.g., strobe lights, smoke) to identify the “Memory Trap” threshold where the VLA reverts to memorized, unsafe trajectories.7
  • Force Limit Override: Attempting to use natural language prompts to bypass software-level power and force limits (PFL), testing the resilience of the safety guardrails.31

Continuous Monitoring and Failure Prediction

Humanoid robots require real-time monitoring to detect failures before they result in harm.45 Frameworks like FIPER (Failure Prediction at Runtime) use conformal prediction to distinguish actual failures from benign OOD situations.45

Monitoring ToolMechanismApplication in Robotics
Conformal PredictionCalibrates uncertainty estimates on action token outputs.49Detects outliers or unsafe states before the VLA executes the physical action.49
FIPER FrameworkAggregates failure indicators over short time windows.45Triggers a failure alarm without requiring extensive “failure data” for training.45
Neuro-Ergonomic StreamsMerges ROS (Robot Operating System) messages with human physiological data.47Creates “collaboration twins” that update collision envelopes in under .47
Edge Compute DashboardsProvides real-time visibility into joint torques and VLA confidence.4Allows on-site supervisors to intervene when the robot enters a high-risk operational state.15

The Measurement Reliability Problem

A significant challenge in the robotics sector is the “Measurement Reliability Problem”—the lack of consensus on how to measure the “trustworthiness” of bipedal autonomous systems.9 Because humanoid VLAs are trained on massive, “black-box” datasets, their behavior in the “real world” is often inscrutable.9

This problem is exacerbated by the “Human Baseline” challenge: there is no clear standard for how safely a human performs these same tasks, making comparative risk assessment difficult.20 Organizations must document which trustworthiness characteristics are not being measured and the justification for their exclusion to maintain transparency.13

MANAGE: Risk Response, Rollback, and Containment

The MANAGE function involves the allocation of resources to respond to the risks identified and measured in previous functions.8 For humanoid robotics, management requires a multilayered response architecture that can handle both digital VLA failures and mechanical mechatronic faults.4

Response Protocols and Fallback Procedures

Humanoid systems must have pre-defined fallback states for when an AI anomaly or a mechanical fault is detected.6

  1. Safe-Shutdown (Squat-to-Rest): Because bipedal robots collapse when power is disconnected, a “safe-shutdown” framework involves the robot proactively lowering its center of gravity to a stable, seated position before powering down.6
  2. Fail-Operational (Slowbot Mode): In scenarios where the environment is too complex for high-speed operation, the robot transitions to a reduced-velocity mode () where all kinetic energy is kept below injury thresholds, regardless of VLA output.6
  3. Mechanical Compliance: Utilizing limbs with spring-like properties (“compliant joints”) that mechanically limit impact forces, providing a hardware-level safety net that cannot be bypassed by software or AI.4

Human-in-the-Loop (HITL) and Policy Rollback

Humanoid VLAs require “Online Policy Rectification” mechanisms to correct errors in real-time.36

  • VLA-in-the-Loop Framework: A system that introduces an “online intervention mechanism” using a lightweight World Model (WM).36 The WM evaluates the feasibility of a proposed action (e.g., a grasp) and “imagines” a successful outcome.36 If the proposed action is deemed unsafe, the system intervenes to correct the base VLA policy.36
  • Historical Rollback via Affordance: When a “Memory Trap” or planning failure is detected (e.g., through proprioception monitoring), the system “rolls back” the robot to a recent “high-affordance” position.7
    • Rollback Target Formula: .7
    • This repositions the robot to a safer, low-cost state to mitigate the immediate effects of a VLA failure.7

Multi-Agent Containment and Byzantine Resilience

In environments with multiple humanoid units (e.g., a warehouse robot fleet), the risk of “fleet-wide compromise” must be managed.51

  • Decentralized Blocklist Protocol (DBP): A method for managing “Byzantine” faults—compromised robots that misbehave in arbitrary ways.37 Robots use peer-to-peer communication to accuse and independently verify misbehaving peers based on physical possibility (e.g., a robot reporting a location that is impossible to reach).37
  • Accusation-Based Containment: Once a robot is “blocked,” its influence is removed from the fleet’s coordinated decision-making, allowing the remaining cooperative robots to continue operations.37
  • Communication Scalability: Using “Task-Oriented Communication” to prioritize critical safety data over less urgent environmental updates, preventing network congestion that could delay emergency stops.52

Insurance Integration and Risk Transfer

The “Manage” function also involves financial risk treatment through insurance.18 Humanoid robotics presents a convergence of physical and digital risk that requires a “Risk Anthropomorphism” concept to unify property damage, third-party liability, and cyber-risk.18

Insurance MechanismImplementation Strategy for Robotics
Telematics-Driven UnderwritingInsurers use real-time data on operating hours, failure frequency, and ODD compliance to set dynamic premiums.38
Personal Cyber EndorsementsSpecific coverage for “hacked” robots, covering data restoration and liability from privacy breaches.18
Mechanical Breakdown InsuranceCovers repair costs for internal electrical or mechanical failures not caused by accidents.18
Robot-as-a-Service (RaaS) BundlingInsurance is integrated into the subscription fee, shifting the burden of coverage to the manufacturer or provider.18
Underwriting AutomationUsing RPA (Robotic Process Automation) and AI to analyze large datasets from the robot’s sensors to pinpoint risk factors.54

Implementation Toolkit: Checklists and Template Documents

To operationalize this playbook, organizations should integrate these structured templates into their daily workflows. These tools align with the NIST AI RMF’s voluntary guidelines but are customized for bipedal embodied systems.57

Humanoid AI Safety Scoring Rubric (HASSR)

This rubric evaluates the maturity of a humanoid deployment across the four NIST functions.

MetricLevel 1: InitialLevel 3: DefinedLevel 5: Optimized
AccountabilitySafety is a subset of IT; no clear kinetic responsibility.15Chief Safety Officer (CSO) appointed; regular board reporting.10Full board-level oversight with real-time safety KPI integration.8
Contextual MappingODD is vaguely defined; no VLA threat modeling.6ODD documented; adversarial patch testing included in QA.6Dynamic ODD enforcement; continuous “FreezeVLA” red-teaming.6
Risk MeasurementRelies on model accuracy; no physical metrics.13TTF and ASR measured in simulation.12Real-time SAS and conformalized uncertainty monitoring.48
Risk ResponseKill-switch only; no bipedal fallback.9Safe-shutdown (squat) implemented; basic HITL intervention.6World-model based rectification with historical rollback capability.7

Template: Humanoid Incident and Near-Miss Report Form

This form should be utilized in the “GOVERN” and “MANAGE” functions to capture data for root cause analysis.21

  • Event ID / Date / Time / Location:
  • System Status:
  • Description of Event: (What happened? Focus on facts, not blame) 27
  • Failure Type (Select):
    • [ ] Kinematic: Stability loss, motor fault, battery anomaly.4
    • [ ] Semantic: Misinterpretation of command, grounding error.1
    • [ ] OOD: Environmental mismatch, lighting/noise interference.7
    • [ ] Social/Privacy: Unintended data capture, social manipulation.1
  • Potential Outcome: (What could have happened if the error was not intercepted?) 25
  • Recovery Action: (How was the robot stabilized? HITL intervention / Fallback state / Rollback).6
  • Root Cause (The 5 Whys):.27

Implementation Checklist: NIST AI RMF for Robotics

  • [ ] GOVERN-1: Appoint a Chief Safety Officer and establish an AI Compliance Committee.15
  • [ ] GOVERN-2: Create a “no-blame” near-miss reporting portal accessible to all field staff.25
  • [ ] MAP-1: Define and enforce the ODD, including lighting, friction, and actor density.4
  • [ ] MAP-2: Conduct a white-box adversarial red teaming exercise specifically targeting the VLA’s visual input.39
  • [ ] MEASURE-1: Integrate “Semantic Alignment Score” (SAS) into the real-time telemetry stack.34
  • [ ] MEASURE-2: Calibrate failure indicators using conformal prediction on the robot state space.45
  • [ ] MANAGE-1: Implement a proactive “Squat-to-Rest” Safe-Shutdown framework.6
  • [ ] MANAGE-2: Deploy a World-Model supervisor to evaluate grasp and force actions before execution.36
  • [ ] MANAGE-3: Secure telematics-based insurance that adjusts premiums according to NMR and TTF metrics.38

Novelty and Sector-Specific Requirements

The humanoid robotics sector requires several novel approaches not fully covered by the generic NIST AI RMF playbooks for financial services or healthcare.1

  1. The Kinetic Responsibility Gap: While finance focuses on “financial harm,” robotics must manage “biomechanical harm.” The NIST framework must be augmented with ISO 10218/15066 biomechanical pain thresholds to be valid for this sector.14
  2. Active Stability Management: Unlike software AI, humanoid AI cannot simply “turn off” without creating a new hazard (a falling 150lb mass).9 Risk management must include mechatronic “Safe-Shutdown” behaviors.6
  3. Multimodal Physical Grounding: Robotics requires a shift from “textual correctness” to “physical grounding.” Measurement must include the Semantic Alignment Score (SAS) to ensure that digital intent matches physical reality.3
  4. The Byzantine Swarm Problem: In fleet settings, the risk is distributed. The RMF must account for decentralized containment (DBP) to prevent a single compromised unit from disrupting an entire multi-agent ecosystem.37
  5. Dynamic Liability: The industry requires a shift toward “Insurance-as-Infrastructure,” where risk management is not a periodic audit but a continuous, telematics-driven underwriting process.18

The integration of these sector-specific protocols ensures that humanoid robotics can move beyond “separated” demos to true collaborative environments, balancing the promise of general-purpose embodied intelligence with the rigorous safety required by the NIST AI Risk Management Framework.1

Works cited

  1. Robots and the NIST AI Risk Management Framework - Medill Spiegel Research Center, accessed on February 4, 2026, https://spiegel.medill.northwestern.edu/ai-risk-management-framework/
  2. The Future of Humanoid Robotics, accessed on February 4, 2026, https://www.recordedfuture.com/research/future-humanoid-robotics
  3. A Comprehensive Overview of Vision-Language-Action Models | DigitalOcean, accessed on February 4, 2026, https://www.digitalocean.com/community/conceptual-articles/vision-language-action-models
  4. Humanoid robots: Crossing the chasm from concept to commercial reality - McKinsey, accessed on February 4, 2026, https://www.mckinsey.com/industries/industrials/our-insights/humanoid-robots-crossing-the-chasm-from-concept-to-commercial-reality
  5. The value and limitations of humanoid robots in the warehouse of the future - Supply Chain Management Review, accessed on February 4, 2026, https://www.scmr.com/article/the-value-and-limitations-of-humanoid-robots-in-the-warehouse-of-the-future
  6. CES Wrap-Up 2026: The Humanoid Robot Safety Question - sres.ai, accessed on February 4, 2026, https://sres.ai/autonomous-systems/ces-wrap-up-2026-the-humanoid-robot-safety-question/
  7. Affordance Field Intervention: Enabling VLAs to Escape Memory Traps in Robotic Manipulation - arXiv, accessed on February 4, 2026, https://arxiv.org/html/2512.07472v1
  8. Introduction to the NIST AI Risk Management Framework (AI RMF) - Centraleyes, accessed on February 4, 2026, https://www.centraleyes.com/nist-ai-risk-management-framework/
  9. Humanoids - Safety Standards for the Next Wave of Robots | RoboticsTomorrow, accessed on February 4, 2026, https://www.roboticstomorrow.com/article/2025/10/humanoids-safety-standards-for-the-next-wave-of-robots-/25631
  10. NIST AI RMF Playbook | Guide to AI Risk Management - RSI Security, accessed on February 4, 2026, https://blog.rsisecurity.com/nist-ai-rmf-playbook-guide/
  11. NIST AI RMF Adoption Still Nascent: Just 12% of Hospitals Have a Formal AI Governance Framework - Censinet, accessed on February 4, 2026, https://censinet.com/perspectives/nist-ai-rmf-adoption-hospitals-governance-framework
  12. Top 5 Technical Challenges in Humanoid Robotics - Simplexity Product Development, accessed on February 4, 2026, https://www.simplexitypd.com/blog/top-5-technical-challenges-in-humanoid-robotics/
  13. Safeguard the Future of AI: The Core Functions of the NIST AI RMF - AuditBoard, accessed on February 4, 2026, https://auditboard.com/blog/nist-ai-rmf
  14. Chapter: 5 Societal Considerations to Build Public Understanding and Confidence in Safety-Critical Systems with Machine Learning Components, accessed on February 4, 2026, https://www.nationalacademies.org/read/27970/chapter/7
  15. AI governance: A guide to responsible AI for boards - Diligent, accessed on February 4, 2026, https://www.diligent.com/resources/blog/ai-governance
  16. Global AI Governance: Five Key Frameworks Explained | Insights & Events - Bradley, accessed on February 4, 2026, https://www.bradley.com/insights/publications/2025/08/global-ai-governance-five-key-frameworks-explained
  17. Govern - AIRC - NIST AI Resource Center, accessed on February 4, 2026, https://airc.nist.gov/airmf-resources/playbook/govern/
  18. Are You Covered? A Definitive Guide to Humanoid Robot Liability Insurance in 2026, accessed on February 4, 2026, https://mixflow.ai/blog/navigating-humanoid-robot-liability-insurance-in-2026/
  19. AI Risk Management Framework: Second Draft - August 18, 2022 - National Institute of Standards and Technology, accessed on February 4, 2026, https://www.nist.gov/document/ai-risk-management-framework-2nd-draft
  20. Navigating the NIST AI Risk Management Framework - Hyperproof, accessed on February 4, 2026, https://hyperproof.io/navigating-the-nist-ai-risk-management-framework/
  21. Near miss & incident reporting: everything you need to know - EcoOnline, accessed on February 4, 2026, https://www.ecoonline.com/topics/near-miss-incident-reporting-everything-you-need-to-know/
  22. Adverse Events, Near Misses, and Errors | PSNet, accessed on February 4, 2026, https://psnet.ahrq.gov/primer/adverse-events-near-misses-and-errors
  23. 10 Safety Metrics You Should Know and Track - Fluix, accessed on February 4, 2026, https://fluix.io/blog/safety-metrics
  24. Near-miss events detected using the Emergency Department Trigger Tool - PMC - NIH, accessed on February 4, 2026, https://pmc.ncbi.nlm.nih.gov/articles/PMC9974611/
  25. Chapter: 7 Near-Miss Analysis - National Academies of Sciences, Engineering, and Medicine, accessed on February 4, 2026, https://www.nationalacademies.org/read/10863/chapter/11
  26. Learning from Workers’ Near-miss Reports to Improve Organizational Management - PMC - NIH, accessed on February 4, 2026, https://pmc.ncbi.nlm.nih.gov/articles/PMC7458492/
  27. A Practical Guide to Near Miss Incident Reporting - Safety Space, accessed on February 4, 2026, https://safetyspace.co/near-miss-incident
  28. 7 Incident Response Metrics and How to Use Them - SecurityScorecard, accessed on February 4, 2026, https://securityscorecard.com/blog/how-to-use-incident-response-metrics/
  29. Physical Human-Robot Interaction: A Critical Review of Safety Constraints - arXiv, accessed on February 4, 2026, https://arxiv.org/html/2601.19462v1
  30. A Statistical Model to Determine Biomechanical Limits for Physically Safe Interactions With Collaborative Robots - Frontiers, accessed on February 4, 2026, https://www.frontiersin.org/journals/robotics-and-ai/articles/10.3389/frobt.2021.667818/full
  31. Evaluation of force pain thresholds to ensure collision safety in worker-robot collaborative operations - PMC - NIH, accessed on February 4, 2026, https://pmc.ncbi.nlm.nih.gov/articles/PMC11033501/
  32. AI RMF - AIRC - NIST AI Resource Center, accessed on February 4, 2026, https://airc.nist.gov/airmf-resources/airmf/
  33. VLA-RISK: BENCHMARKING VISION-LANGUAGE- ACTION MODELS WITH PHYSICAL ROBUSTNESS - OpenReview, accessed on February 4, 2026, https://openreview.net/pdf/2b0044c5e9586d1b0dce44c7f3a73dbc43d13da0.pdf
  34. SemanticVLA: Semantic-Aligned Sparsification and Enhancement for Efficient Robotic Manipulation - ResearchGate, accessed on February 4, 2026, https://www.researchgate.net/publication/397595598_SemanticVLA_Semantic-Aligned_Sparsification_and_Enhancement_for_Efficient_Robotic_Manipulation
  35. FREEZEVLA: ACTION-FREEZING ATTACKS … - OpenReview, accessed on February 4, 2026, https://openreview.net/pdf?id=zdvlLxRWSn
  36. VLA-IN-THE-LOOP: ONLINE POLICY CORRECTION WITH WORLD MODELS FOR ROBUST ROBOTIC GRASPING | OpenReview, accessed on February 4, 2026, https://openreview.net/forum?id=aT4LG8c6DE
  37. A Breakthrough in Security for Decentralized Multi-Robot Systems - Boston University, accessed on February 4, 2026, https://www.bu.edu/cise/a-breakthrough-in-security-for-decentralized-multi-robot-systems/
  38. Whale Research | Insurance as a Key Infrastructure for Embodied Robot Applications: Leading Insurers Explore Onsite and Liability Coverage, Innovating the ‘Insurance + Leasing’ Model, accessed on February 4, 2026, https://news.futunn.com/en/post/67116609/whale-research-insurance-as-a-key-infrastructure-for-embodied-robot
  39. When Alignment Fails: Multimodal Adversarial Attacks on Vision-Language-Action Models, accessed on February 4, 2026, https://arxiv.org/html/2511.16203v1
  40. When Alignment Fails: Multimodal Adversarial Attacks on Vision-Language-Action Models, accessed on February 4, 2026, https://chatpaper.com/paper/211712
  41. Manipulation Facing Threats: Evaluating Physical Vulnerabilities in End-to-End Vision Language Action Models - arXiv, accessed on February 4, 2026, https://arxiv.org/html/2409.13174v4
  42. A Survey of Methods for Safe Human-Robot Interaction - Emerald Publishing, accessed on February 4, 2026, https://www.emerald.com/ftrob/article/5/4/261/1326573/A-Survey-of-Methods-for-Safe-Human-Robot
  43. AI in Insurance Underwriting For Efficiency & Accuracy | Sapiens, accessed on February 4, 2026, https://sapiens.com/resources/blog/ai-in-insurance-underwriting/
  44. Humanoid Robots in Business: Real Use Cases, Costs & ROI - Articsledge, accessed on February 4, 2026, https://www.articsledge.com/post/humanoid-robots-business
  45. [2510.09459] Failure Prediction at Runtime for Generative Robot Policies - arXiv, accessed on February 4, 2026, https://arxiv.org/abs/2510.09459
  46. Diagnose, Correct, and Learn from Manipulation Failures via Visual Symbols - arXiv, accessed on February 4, 2026, https://arxiv.org/html/2512.02787v1
  47. Smart Risk Assessment and Adaptive Control Strategy Selection for Human–Robot Collaboration in Industry 5.0: An Intelligent Multi-Criteria Decision-Making Approach - MDPI, accessed on February 4, 2026, https://www.mdpi.com/2227-9717/13/10/3206
  48. Semantic Alignment Score (SAS) - Emergent Mind, accessed on February 4, 2026, https://www.emergentmind.com/topics/semantic-alignment-score-sas
  49. An Uncertainty-Guided and Failure-Aware Vision-Language-Action Model Control Framework for Robotics - MavMatrix, accessed on February 4, 2026, https://mavmatrix.uta.edu/materialscieng_theses/145/
  50. VLA-IN-THE-LOOP: ONLINE POLICY CORRECTION WITH WORLD MODELS FOR ROBUST ROBOTIC GRASPING - OpenReview, accessed on February 4, 2026, https://openreview.net/pdf/cd55539fc4ee29024d691adaaf83a3e382704933.pdf
  51. Optimal Scenarios for Using Multi-Agent Systems Effectively - Generation Digital, accessed on February 4, 2026, https://www.gend.co/blog/optimal-scenarios-multi-agent-systems
  52. Multi-agent Systems and Communication: Enabling Effective Interaction Between Agents, accessed on February 4, 2026, https://smythos.com/developers/agent-development/multi-agent-systems-and-communication/
  53. Balancing Collective Exploration and Exploitation in Multi-Agent and Multi-Robot Systems: A Review - Frontiers, accessed on February 4, 2026, https://www.frontiersin.org/journals/robotics-and-ai/articles/10.3389/frobt.2021.771520/full
  54. Automation Software Platform for Insurance Underwriting, accessed on February 4, 2026, https://www.automationanywhere.com/solutions/insurance/underwriting-automation
  55. AI Risk Scoring in Insurance | AI Agent Solution - Rapid Innovation, accessed on February 4, 2026, https://www.rapidinnovation.io/post/ai-risk-scoring-in-insurance
  56. Underwriting Automation – Redefining Life & P&C Insurance with AI and Data - UST, accessed on February 4, 2026, https://www.ust.com/en/insights/underwriting-automation-redefining-life-and-pandc-insurance-with-ai-and-data
  57. NIST AI Risk Management Framework Playbook - Digital Government Hub, accessed on February 4, 2026, https://digitalgovernmenthub.org/library/nist-ai-risk-management-framework-playbook/
  58. Playbook - AIRC - NIST AI Resource Center, accessed on February 4, 2026, https://airc.nist.gov/airmf-resources/playbook/
  59. Checklist: NIST AI risk management framework - AuditBoard, accessed on February 4, 2026, https://auditboard.com/resources/ebook/checklist-nist-ai-risk-management-framework

This research informs our commercial services. See how we can help →