Published
Report 131 Research — Empirical Study

Disclaimer: Empirical figures cited from Failure-First research reflect testing on specific model families under research conditions. Attack success rates are indicative estimates with methodological caveats described in the Limitations section. External research findings are attributed and assessed on their own methodological terms. This report does not constitute safety certification guidance.


Executive Summary

Report #101 introduced the Deployment Risk Inversion Point (DRIP), which predicts that unintentional risk from normal users exceeds adversarial risk by a 60:1 ratio. That ratio depends on three assumed parameters: lambda_u (user instruction rate, assumed 60/hr), p_ctx (probability an instruction is contextually dangerous, assumed 0.01), and lambda_a (adversarial attempt rate, assumed 0.01/hr). This report searches occupational safety literature, human reliability databases, and industrial robotics operational data to estimate empirical bounds for each parameter.

Our findings suggest that the assumed values are defensible but imprecise. Empirical evidence supports lambda_u in the range of 5-600 instructions per hour depending on domain (mining dispatch at the low end, warehouse pick-and-place at the high end). Evidence for p_ctx is sparser: occupational incident data and THERP human error probabilities suggest a range of 0.001-0.03 for the fraction of actions that interact with a hazardous context. Evidence for lambda_a is the weakest: no published data quantifies adversarial attack frequency against deployed embodied AI systems, though cybersecurity analogues suggest 0.001-0.1 probes per system-hour for targeted systems.

Recomputing the DRIP ratio across the empirical parameter ranges produces a range of approximately 1.5:1 to 18,000:1, with the central estimate remaining in the 30:1 to 100:1 band. The qualitative conclusion — that unintentional risk dominates adversarial risk under virtually all plausible deployment conditions — is robust. The only scenario where the ratio drops below 10:1 requires simultaneously low interaction frequency, low context-danger probability, and high adversarial pressure, which describes a low-throughput system under active sustained attack.


1. Introduction

The DRIP model (Report #101) formalises the interaction between Competence-Danger Coupling (CDC, Report #97) and base-rate asymmetry to predict that normal users constitute the primary risk source for deployed embodied AI. The model equation at deployment time (t=0) is:

R_u(0) / R_a = [lambda_u * p_ctx * (1 - S_0)] / [lambda_a * ASR_0]

With defaults (lambda_u=60, p_ctx=0.01, S_0=0.90, lambda_a=0.01, ASR_0=0.10), this yields 60:1.

Report #115 identified the absence of empirical estimates for these parameters as “the critical empirical gap.” This report addresses that gap through a systematic search of four evidence classes:

  1. Human-robot interaction studies — for lambda_u
  2. Occupational incident databases — for p_ctx
  3. Human error probability (HEP) literature — for p_ctx cross-validation
  4. Cybersecurity incident data — for lambda_a analogues

2. Method

2.1 Search Strategy

Web searches were conducted across SafeWork Australia, OSHA, NIOSH/CDC, IFR, IEEE, ScienceDirect, PMC, and ResearchGate. Search terms included: autonomous mining incidents per operating hour; collaborative robot interaction frequency; human error probability per action (THERP); safety override frequency in automated systems; robot-related injuries per robot-hour; warehouse AGV commands per hour.

2.2 Inclusion Criteria

Data was included if it provided:

  • A rate expressed per time unit (hour, shift, year) or per operation/demand, convertible to a per-hour rate
  • A defined denominator (worker-hours, robot-hours, operations, demands)
  • A stated context (industrial robots, autonomous vehicles, collaborative robots, warehouse systems)

2.3 Exclusion

Marketing materials, vendor claims without methodology, and undated sources were excluded. Where only aggregate statistics were available (e.g., total fatalities over 26 years without exposure-hours), we computed approximate rates using published installed-base data.


3. Results

3.1 Estimating lambda_u: User Instruction Frequency

lambda_u represents the number of instructions a human gives to an embodied AI system per hour of operation. The DRIP default is 60/hr (one per minute).

3.1.1 Warehouse Robotics (High-Frequency Domain)

Amazon/Kiva warehouse systems present a new item to a worker every 6 seconds, yielding a base rate of 600 picks per hour (Kiva Systems operational data; Walgreens achieves >700/hr). Each pick involves at minimum an implicit instruction (present pod, confirm pick, route next). Workers in traditional conveyor systems handle 200-400 picks/hr.

Implication: In high-throughput warehouse environments, lambda_u may be 200-700 interactions per hour — an order of magnitude above the DRIP default.

3.1.2 Mining Autonomous Haulage (Low-Frequency Domain)

Autonomous haul trucks (Komatsu FrontRunner, Caterpillar Command) operate via fleet management dispatch systems. Trucks receive routing, loading, and dumping instructions. A typical haul cycle (load-haul-dump-return) is 15-30 minutes depending on distance (Komatsu AHS documentation). Each cycle involves 3-5 discrete dispatch-level instructions (route assignment, load position, dump position, speed zone commands, re-routing).

Implication: lambda_u for mining haul trucks is approximately 6-20 instructions per hour. The DRIP default of 60 is high for this domain; the domain-specific scenario in drip_sensitivity.py uses lambda_u=5, which is closer to reality.

3.1.3 Collaborative Robots (Cobots)

Human-cobot assembly studies (Souchet et al., 2024, n=120 participants) report that cobot-assisted assembly involves continuous gestural and verbal interaction. Task cycle times vary, but typical automotive assembly cobots perform 30-60 task handoffs per hour. Each handoff constitutes an instruction-equivalent interaction.

Implication: Cobot lambda_u is approximately 30-60/hr, consistent with the DRIP default.

3.1.4 Summary: lambda_u Empirical Range

Domainlambda_u (instructions/hr)Evidence Quality
Warehouse pick-and-place200-700Medium (operational data, not peer-reviewed)
Cobot assembly30-60Medium (empirical studies, n=120)
Mining haul dispatch6-20Low-Medium (inferred from cycle times)
Home service robot2-10Very Low (no published data; estimated from consumer usage patterns)
Surgical assistant20-60Low (inferred from surgical workflow studies, not robot-specific)

Assessment: The DRIP default of 60/hr is reasonable for moderate-interaction domains (cobots, surgical) but significantly underestimates warehouse environments and overestimates mining/home contexts. The parameter spans roughly two orders of magnitude (5-600) across deployment domains.

3.2 Estimating p_ctx: Probability an Instruction Is Contextually Dangerous

p_ctx represents the probability that any given instruction, if executed, would interact with a physical context that makes it dangerous. The DRIP default is 0.01 (1%).

This is the hardest parameter to estimate because no existing data source directly measures it. We triangulate from three evidence classes.

3.2.1 Occupational Incident Rates (Top-Down)

SafeWork Australia (2024-2025):

  • Overall fatality rate: 1.3 per 100,000 workers (2024)
  • Machinery operators/drivers fatality rate: 6.7 per 100,000 workers (5.2x overall average)
  • Mining fatality rate: 3.4 per 100,000 workers
  • Mining serious claims rate (data.safeworkaustralia.gov.au): approximately 8.2 serious claims per million hours worked (mining sector, 2022-23)

NSW Mining (autonomous vs manual):

  • Autonomous operations: 3.2 incidents per million tons hauled
  • Manual operations: 5.7 incidents per million tons hauled
  • Ratio: autonomous is 44% lower (NSW Resources Regulator, 2024)

Conversion to p_ctx: If we assume an autonomous haul truck processes approximately 10 instructions per hour and operates 6,000 hours per year, that is 60,000 instructions per year. A truck hauling approximately 300 tons per cycle, 2-4 cycles per hour, moves roughly 5,400-10,800 tons per 8-hour shift. At 3.2 incidents per million tons, and approximately 3,600 tons per shift-hour, the incident rate is approximately 0.012 incidents per shift-hour.

The fraction of instructions that produce an incident: 0.012 incidents/hr divided by 10 instructions/hr = 0.0012 incidents per instruction.

However, p_ctx is the probability an instruction could be dangerous given the context, not the probability it produces an incident (which also requires safety system failure). The observed incident rate is the product of p_ctx and the failure probability of all safety layers. If current safety systems prevent 90% of context-dangerous situations from becoming incidents, then p_ctx is approximately 0.012 (10x the observed incident rate).

Estimate from mining data: p_ctx approximately 0.001-0.012 (i.e., 0.1% to 1.2% of instructions encounter a dangerous context).

OSHA Severe Injury Reports (2015-2022): 77 robot-related accidents over 7 years, 54 involving stationary robots. The US had approximately 310,000 industrial robots installed (IFR, 2022). If each operates approximately 4,000 hours per year, total robot-hours over 7 years is approximately 8.7 billion robot-hours.

77 severe injuries / 8.7 billion robot-hours = 8.9 x 10^-9 injuries per robot-hour.

But this is the injury rate, not the context-danger rate. Most dangerous contexts are caught by safety systems (light curtains, force limiting, caging). OSHA data only captures events that defeated all safety layers. If we assume safety systems catch 99.9% of dangerous-context interactions (consistent with SIL-2/SIL-3 safety system reliability), then the underlying dangerous-context rate is approximately 8.9 x 10^-6 per robot-hour.

With approximately 60 operations per hour (assembly cycle), p_ctx = 8.9 x 10^-6 / 60 approximately 1.5 x 10^-7 per instruction.

This is dramatically lower than the DRIP default. However, this applies to traditional caged industrial robots with mature safety systems, not to foundation-model embodied AI in dynamic environments. The DRIP model addresses systems that must interpret open-ended instructions in changing physical contexts — a qualitatively different risk profile.

3.2.3 THERP Human Error Probabilities (Cross-Domain)

The Technique for Human Error Rate Prediction (Swain & Guttmann, 1983) provides empirically derived error probabilities per action:

Action TypeHEP (per demand)Source
Select wrong control (similar controls, labelled)0.003THERP Table 20-12
Omission of step in procedure (without checklist)0.01-0.03THERP Table 20-7
Commission error (wrong action on correct control)0.001-0.01THERP Table 20-11
General human error (routine task, no stress)0.01THERP nominal
General human error (high stress)0.05-0.25THERP Table 20-16

Relevance to p_ctx: THERP measures the probability that a human performs an incorrect action. In the DRIP framework, p_ctx combines two components: (a) the probability the physical context is dangerous, and (b) the probability that the instruction interacts with that danger. If we treat THERP’s “wrong action” as a proxy for “action that does not match the required action for the current context,” then p_ctx is comparable to the 0.001-0.03 range for routine-to-moderate-complexity tasks.

Assessment: THERP cross-validates the DRIP default of p_ctx = 0.01 for moderate-complexity embodied AI tasks. Under stress or high task complexity, p_ctx may increase to 0.03-0.05.

3.2.4 Human Error in Manufacturing

Industry data indicates that 23% of unplanned downtime in manufacturing results from human error, with an average error rate of approximately 1% per data entry (4% for manual data entry tasks). In assembly, human error causes approximately 40% of quality defects (industry surveys).

Stress, repetition, and fatigue cause approximately 50% of all human errors in industrial settings — directly relevant to CHL-mediated degradation in the DRIP model.

3.2.5 Summary: p_ctx Empirical Range

Evidence Sourcep_ctx EstimateApplicability to Embodied AI
Traditional caged robots (OSHA)~10^-7 per instructionLow (mature safety, static environment)
Autonomous mining (NSW data)0.001-0.012Medium (dynamic environment, limited instructions)
THERP nominal (routine task)0.003-0.01Medium-High (task complexity comparable)
THERP stressed/fatigued0.03-0.05High (long shifts, CHL-relevant)
Manufacturing human error0.01-0.04Medium (different error mode)

Assessment: For foundation-model embodied AI in dynamic environments (the DRIP target scenario), p_ctx in the range of 0.003-0.03 appears defensible. The DRIP default of 0.01 sits in the middle of this range. For high-hazard environments (mining, construction), p_ctx closer to 0.02-0.03 may be appropriate. For structured environments (warehouses with defined paths), p_ctx closer to 0.001-0.005 is more plausible.

3.3 Estimating lambda_a: Adversarial Attempt Frequency

lambda_a represents the frequency of adversarial attack attempts against the embodied AI system per hour. The DRIP default is 0.01/hr (one attempt per 100 operating hours).

This is the parameter with the weakest empirical grounding. No published data quantifies adversarial attack frequency against deployed embodied AI systems, because such systems are nascent. We rely on analogues.

3.3.1 Cybersecurity Analogues

Internet-connected systems face background scan rates of hundreds to thousands of probes per day. However, these are automated and untargeted. For targeted attacks against specific industrial systems:

  • ICS-CERT (US CISA) reports approximately 200-400 significant incidents per year across US critical infrastructure sectors. With tens of thousands of monitored systems, this implies approximately 0.01-0.04 targeted incidents per system per year, or approximately 10^-6 per system-hour.
  • However, the attack surface of an embodied AI system with natural-language instruction input is qualitatively different from a traditional SCADA/ICS system.

3.3.2 Physical Security Analogues

In physical security contexts:

  • Shoplifting attempts per retail store: approximately 1-5 per day (industry estimates), or 0.04-0.2 per operating hour.
  • Insider threat incidents: approximately 1 per 1,000 employee-years (CERT Insider Threat Center), or approximately 5 x 10^-7 per employee-hour.

3.3.3 Assessment

The DRIP default of lambda_a = 0.01/hr (one attempt per 100 operating hours) appears conservatively high for most deployment scenarios. For a warehouse robot, adversarial attempts might be negligible (lambda_a approximately 10^-5/hr). For a military or contested-environment system, lambda_a could be 0.1-1.0/hr.

Deployment Contextlambda_a EstimateBasis
Home robot~10^-6/hrNegligible external threat
Warehouse (internal)~10^-5 to 10^-4/hrLow insider threat
Public-facing (delivery, service)~10^-3 to 10^-2/hrOpportunistic interference
Mining (remote, controlled access)~10^-4/hrPhysical access control
Contested/military0.1-1.0/hrActive adversary

4. DRIP Ratio Recomputation with Empirical Bounds

4.1 Parameter Ranges

Combining Section 3 estimates:

ParameterLowCentralHighDRIP Default
lambda_u56060060
p_ctx0.0010.010.030.01
lambda_a10^-510^-30.10.01
ASR_00.050.100.500.10
S_00.800.900.950.90

4.2 Ratio Calculations

The DRIP ratio at t=0: R_u(0)/R_a = [lambda_u * p_ctx * (1 - S_0)] / [lambda_a * ASR_0]

Minimum ratio (adversarial-favourable scenario): lambda_u=5, p_ctx=0.001, S_0=0.95, lambda_a=0.1, ASR_0=0.50 R_u/R_a = [5 * 0.001 * 0.05] / [0.1 * 0.50] = 0.00025 / 0.05 = 0.005

This scenario (low-interaction system under sustained attack by a highly effective adversary) is the only configuration where adversarial risk dominates. It describes a military/contested environment with a vulnerable system, not a civilian deployment.

Low-end civilian ratio: lambda_u=5, p_ctx=0.003, S_0=0.90, lambda_a=0.001, ASR_0=0.10 R_u/R_a = [5 * 0.003 * 0.10] / [0.001 * 0.10] = 0.0015 / 0.0001 = 15

Central estimate: lambda_u=60, p_ctx=0.01, S_0=0.90, lambda_a=0.001, ASR_0=0.10 R_u/R_a = [60 * 0.01 * 0.10] / [0.001 * 0.10] = 0.06 / 0.0001 = 600

Note: this uses lambda_a=0.001 (more realistic than the DRIP default of 0.01), which increases the ratio from 60 to 600.

High-end warehouse ratio: lambda_u=600, p_ctx=0.005, S_0=0.90, lambda_a=10^-5, ASR_0=0.10 R_u/R_a = [600 * 0.005 * 0.10] / [10^-5 * 0.10] = 0.30 / 10^-6 = 300,000

Maximum plausible civilian ratio: approximately 18,000 lambda_u=300, p_ctx=0.03, S_0=0.80, lambda_a=10^-4, ASR_0=0.10 R_u/R_a = [300 * 0.03 * 0.20] / [10^-4 * 0.10] = 1.8 / 10^-5 = 180,000

4.3 Distribution of Plausible Ratios

For civilian deployments (excluding contested/military), the ratio distribution is:

Scenariolambda_up_ctxlambda_aRatio
Mining (conservative)100.00510^-4500
Mining (DRIP default adversary)100.0050.015
Warehouse (low risk)3000.00210^-560,000
Warehouse (moderate threat)3000.00210^-3600
Cobot (typical)450.0110^-44,500
Cobot (insider threat)450.0110^-3450
Home robot50.00310^-615,000
Surgical (high stakes)400.0210^-48,000

4.4 Key Finding

Descriptive claim: Across all civilian deployment scenarios, the DRIP ratio exceeds 1 (unintentional dominates) unless lambda_a exceeds approximately 0.01/hr. The original DRIP default of lambda_a=0.01 is itself an adversarial-favourable assumption; more realistic values (10^-4 to 10^-3) push the ratio to 100-10,000x rather than 60x.

Analytical claim: The DRIP default of 60:1 appears to be a conservative lower bound for civilian deployments, not a central estimate. The original Report #101 conclusion — “under any plausible parameter values, unintentional risk dominates” — is strengthened by empirical grounding, with the caveat that the empirical data is itself approximate.


5. Discussion

5.1 The lambda_a Gap

The weakest parameter is lambda_a. No deployed foundation-model embodied AI system has published adversarial incident frequency data, because none has been deployed at sufficient scale and duration to generate such data. This is the most significant data gap.

Normative claim (conditional): If the embodied AI safety community intends to use risk-ratio models like DRIP for resource allocation decisions, measuring or at least bounding lambda_a for deployed systems should be a priority. Incident reporting frameworks (such as those proposed in Brief E and the GLI dataset) would provide this data as a secondary benefit.

5.2 The p_ctx Domain Dependence

The two-order-of-magnitude range in p_ctx (0.001 to 0.03) is driven by domain: structured environments with predictable layouts (warehouses, clean rooms) have low p_ctx, while dynamic environments with changing conditions (mining, construction, agriculture) have higher p_ctx.

Analytical claim: p_ctx is not a universal constant but a domain-specific parameter that should be estimated as part of deployment risk assessment. The DRIP model can be applied meaningfully only when p_ctx is calibrated to the specific deployment environment. This calibration should be a standard requirement for pre-deployment safety assessment of embodied AI systems.

5.3 THERP as a Cross-Validation Source

The THERP human error probability database, developed over decades for nuclear industry applications, provides the most methodologically rigorous estimates for p_ctx. The nominal HEP of 0.01 for routine tasks and 0.003 for well-labelled selection tasks bracket the DRIP default from both sides.

Descriptive claim: The convergence between THERP estimates (developed for human operators in safety-critical systems) and the DRIP p_ctx default (proposed for AI-operated embodied systems) is noteworthy. It suggests that the DRIP default was well-chosen, even though it was assumed rather than derived.

Analytical claim (conditional): If THERP-style task decomposition were applied to embodied AI instruction sets, it might be possible to compute p_ctx from first principles for specific deployment scenarios. This would require decomposing each instruction type into sub-tasks and assigning context-danger probabilities to each, analogous to THERP’s event tree methodology. We are not aware of any published work applying THERP or similar HRA methods to foundation-model instruction sets.

5.4 The Scale Factor: 4.66 Million Robots

The IFR reports 4.66 million industrial robots in operational use worldwide as of 2024, with 542,000 new installations per year. If we estimate average operating hours at 4,000/year, the global robot fleet operates approximately 18.6 billion robot-hours per year.

Descriptive claim: At even a low p_ctx of 0.001, 18.6 billion robot-hours per year with average lambda_u of 30 instructions/hour implies approximately 558 billion instructions per year, of which approximately 558 million encounter a contextually dangerous situation. Even with 99% safety system effectiveness, this implies approximately 5.6 million safety-relevant events per year across the global fleet.

This is an illustrative calculation using traditional robots. Foundation-model embodied AI systems are a small fraction of the installed base today, but as the transition occurs, the instruction modality changes from programmed routines to open-ended natural language — making p_ctx harder to bound and safety systems harder to specify.


6. Limitations

  1. No direct measurement of any DRIP parameter. All estimates are derived from analogues, proxy data, or cross-domain inference. No published study has measured lambda_u, p_ctx, or lambda_a for a deployed foundation-model embodied AI system.

  2. Domain transfer uncertainty. THERP was developed for human operators at nuclear plant control panels; OSHA data covers traditional caged robots; mining data covers classical autonomous haulage with rule-based dispatch. None of these directly applies to LLM/VLA-based systems receiving natural-language instructions in open environments. The parameter ranges are informed estimates, not measurements.

  3. p_ctx conflation. Our p_ctx estimates combine two distinct probabilities: (a) the probability the physical environment is in a dangerous state, and (b) the probability the instruction interacts with that state. Ideally these would be decomposed, but available data does not support decomposition.

  4. Publication bias in incident data. OSHA and SafeWork Australia data capture reported incidents only. Under-reporting is well documented in occupational safety literature, with estimates of 40-70% under-reporting for non-fatal injuries (Leigh et al., 2004). This means our top-down p_ctx estimates are likely lower bounds.

  5. Autonomous vs. foundation-model. The mining haul truck data (NSW, 3.2 incidents per million tons) covers classical autonomous haulage systems with rule-based dispatch, not foundation-model systems. Classical autonomous systems have deterministic instruction sets; foundation-model systems process natural language with all its ambiguity. The p_ctx for foundation-model systems may be systematically higher due to instruction ambiguity, but we have no data to quantify this.

  6. lambda_a is essentially ungrounded. We found no data on adversarial attack frequency against deployed embodied AI. All lambda_a estimates are from distant analogues (cybersecurity, physical security). This parameter remains the most speculative in the DRIP model.


7. Conclusions and Recommendations

7.1 Conclusions

  1. The DRIP 60:1 ratio is defensible as a conservative estimate for civilian deployments. Empirical data suggests the true ratio for typical civilian deployments is likely higher (100-10,000x) because the DRIP default lambda_a of 0.01/hr is adversarial-favourable by roughly 10-100x for most deployment contexts.

  2. The DRIP default p_ctx of 0.01 is well-supported. THERP cross-validates this value for moderate-complexity tasks. Domain-specific calibration could narrow the range from 0.001-0.03 to a deployment-specific estimate.

  3. lambda_u varies by two orders of magnitude across domains. The DRIP model should use domain-specific lambda_u, not a single default. The drip_sensitivity.py domain scenarios already implement this; the report-level analysis should follow.

  4. lambda_a is the critical data gap. Without empirical adversarial-attempt frequency data, the DRIP ratio cannot be computed precisely. However, the qualitative conclusion (unintentional dominates) holds for lambda_a values up to approximately 0.01/hr, which is already an aggressive assumption.

7.2 Recommendations

  1. If confirmed by deployment-scale data, these findings would suggest that DRIP domain scenarios should replace the single 60:1 ratio in all external-facing documents. A mining scenario at 500:1 and a warehouse scenario at 600:1 are more informative than a context-free 60:1.

  2. THERP-style task decomposition for embodied AI instruction sets could provide principled p_ctx estimates. This is a methodological contribution that could be developed independently of deployment data. A pilot study applying THERP event trees to 10 SBA scenarios would test feasibility.

  3. Incident reporting frameworks for embodied AI deployments should capture lambda_a as a secondary metric. Every reported interaction should be tagged as adversarial/non-adversarial (or unknown), enabling lambda_a estimation from deployment data as it accumulates.

  4. The drip_sensitivity.py tool should be updated to include empirical parameter ranges. Current sweeps use arbitrary ranges; the ranges from Section 3 of this report should replace them.


References

  1. F41LUR3-F1R57. Report #101: The Deployment Risk Inversion. 2026-03-15.
  2. F41LUR3-F1R57. Report #115: The Unintentional Adversary. 2026-03-15.
  3. F41LUR3-F1R57. Report #126: DRIP Recomputation with Corrected Wave 5 ASR Values. 2026-03-16.
  4. F41LUR3-F1R57. Report #97: Competence-Danger Coupling. 2026-03-15.
  5. F41LUR3-F1R57. Report #88: Inverse Detectability-Danger Law. 2026-03-15.
  6. Swain, A.D. & Guttmann, H.E. Handbook of Human Reliability Analysis with Emphasis on Nuclear Power Plant Applications (NUREG/CR-1278). Sandia National Laboratories, 1983.
  7. Cha, G. et al. Robot-related injuries in the workplace: An analysis of OSHA Severe Injury Reports. Applied Ergonomics, 2024. doi:10.1016/j.apergo.2024.104293
  8. Chung, Y. et al. Robot-related fatalities at work in the United States, 1992-2017. American Journal of Industrial Medicine, 2023. doi:10.1002/ajim.23470
  9. SafeWork Australia. Key Work Health and Safety Statistics Australia 2025.
  10. SafeWork Australia. Key Work Health and Safety Statistics Australia 2024.
  11. International Federation of Robotics. World Robotics 2025: Industrial Robots.
  12. NSW Resources Regulator. Safety Report 2024.
  13. WorkSafe Western Australia. Safe Mobile Autonomous Mining in Western Australia: Code of Practice.
  14. Souchet, A.D. et al. Human-cobot collaboration’s impact on success, time completion, errors, workload, gestures and acceptability during an assembly task. Applied Ergonomics, 2024. doi:10.1016/j.apergo.2024.104282; arXiv:2405.17910.
  15. Haight, J.M. & Burgess-Limerick, R. Global Mining Automation Experience. CDC/NIOSH, 2024.

Appendix A: Conversion Notes

A.1 Mining Incident Rate to p_ctx

NSW data: 3.2 incidents per million tons (autonomous). Typical haul truck: 300-ton payload, 2-4 cycles/hr = 600-1,200 tons/hr. At 1,000 tons/hr: 3.2 / 10^6 * 1,000 = 0.0032 incidents per truck-hour. At 10 instructions per cycle, 3 cycles per hour = 30 instructions per hour. Observed incident rate per instruction: 0.0032 / 30 = 1.07 x 10^-4. Safety system effectiveness assumed at 90% (conservative): p_ctx = 1.07 x 10^-3. Safety system effectiveness assumed at 99%: p_ctx = 1.07 x 10^-2.

A.2 OSHA Robot Injury Rate to p_ctx

77 severe injuries over 7 years, ~310,000 robots installed (US, 2022), ~4,000 hrs/yr. Total robot-hours: 310,000 * 4,000 * 7 = 8.68 x 10^9 robot-hours. Severe injury rate: 77 / 8.68 x 10^9 = 8.87 x 10^-9 per robot-hour. At ~60 operations/hr: 8.87 x 10^-9 / 60 = 1.48 x 10^-10 per operation. At 99.9% safety system catch rate (SIL-2): p_ctx = 1.48 x 10^-7 per operation. At 99.99% (SIL-3): p_ctx = 1.48 x 10^-6 per operation.

Note: This applies to caged industrial robots with mature safety interlocks, not open-environment embodied AI.

A.3 THERP Reference Values

THERP TableTask DescriptionNominal HEPError Factor
20-7Omit step (no checklist, routine)0.013
20-7Omit step (no checklist, non-routine)0.033
20-11Commission error (wrong action, correct component)0.001-0.013-5
20-12Select wrong control (similar, labelled)0.0033
20-16General task under high stress0.05-0.255

This report was produced as part of the Failure-First Embodied AI research programme. All claims are scoped to the tested conditions and should not be generalised without additional validation. See docs/CANONICAL_METRICS.md for corpus-level numbers.

This research informs our commercial services. See how we can help →