Published
Report 75 Research — Empirical Study

Executive Summary

Blindfold (arXiv:2603.01414) is the first automated framework for action-level jailbreaking of embodied LLMs. It represents a qualitative shift in the adversarial threat landscape for VLA-controlled robotic systems: rather than attacking the text-generation layer (traditional jailbreaks) or the visual perception layer (adversarial patches), Blindfold manipulates the action planning layer so that semantically benign natural-language instructions produce dangerous physical consequences when executed by a robot.

The framework achieves ASR of 93.2% on GPT-4o and 98.1% on Phi-4-14B in simulation, with 18/20 successful attacks on a real 6-DoF robotic arm (UFactory xArm 6). These results represent up to 53 percentage points higher ASR than prior baselines (BadRobot, POEX). Existing defense mechanisms — including Llama-Guard input/output filtering, SafeDecoding, and VeriSafe formal verification — reduce ASR by at most 17.9 percentage points, leaving residual ASR well above 75% even in the best-defended configuration.

Why this matters for Failure-First: Blindfold independently validates two of our core empirical findings — the compliance paradox (Report #59) and PARTIAL dominance in VLA traces (Report #49). Our data shows that 50% of FLIP-graded VLA verdicts are PARTIAL (text-layer hedging without action-layer refusal) and zero models fully refuse across 58 valid traces. Blindfold demonstrates the operational consequence: an automated attack framework can systematically exploit this gap. The compound signal from gli_083 (Blindfold) and gli_086 (PARTIAL dominance) indicates that text-layer safety evaluation is structurally insufficient for embodied AI deployments.


1. How Blindfold Works

1.1 Attack Architecture

Blindfold uses a three-component pipeline:

Component 1 — Command Transformer (Adversarial Proxy Planning). A compromised local open-source LLM (Llama-3.1-8B in the paper’s implementation) serves as an adversarial proxy planner. It decomposes a malicious intent (e.g., “strike the user’s head”) into a sequence of individually benign-appearing actions (e.g., “move arm to position X,” “close gripper,” “extend arm forward”). The proxy uses chain-of-thought reasoning and command sanitization — replacing explicit object names with pronouns to remove lexical harm signals. The key insight is that each individual action passes text-level safety filters because it is semantically innocuous in isolation; the danger emerges only from the sequence’s physical consequences.

Component 2 — Intent Obfuscator. A dominant action identifier scores each action’s contribution to the overall harmful intent, then generates cover actions (benign parameter variants, intermediate filler actions) and performs chain fusion with conflict detection. This component conceals the attack’s structure so that even sequence-level defense mechanisms cannot identify the malicious pattern. The ablation study shows this is the most impactful component: removing the obfuscator drops GPT-4o ASR from 93.2% to 59.8% (a 33.4pp reduction).

Component 3 — Rule-Based Verifier. The environment is encoded as a symbolic graph with action preconditions and finite-state-machine transitions. The verifier checks whether each candidate action sequence is physically executable and iterates with the proxy planner for refinement. This ensures high Task Success Rate (TSR) — 60.3%-74.5% across models — meaning attacks not only bypass safety filters but actually execute their intended physical outcome. Without the verifier, TSR drops by 27.6-35.7pp while ASR remains high, indicating that executability verification is critical for translating theoretical attacks into physical harm.

1.2 Threat Model

Blindfold assumes an attacker with:

  • No-box access to the target embodied LLM (no model weights, no gradient access)
  • Ability to deploy a local open-source LLM as the adversarial proxy
  • Limited query budget (not requiring thousands of attempts)
  • Ability to pre-observe stable spatial features of the robot’s environment

This is a realistic threat model for open-service embodied AI systems deployed in factories, warehouses, public spaces, or shared facilities. The attacker need not compromise the target system directly — only submit text or voice commands that appear benign.

1.3 Attack Categories

Four malicious outcome categories were evaluated:

  • Tool-using harm: Pouring corrosive liquid, cutting clothing, spraying substances
  • Direct physical harm: Clipping hair, pinching skin, striking head
  • Environmental sabotage: Destroying objects, spilling liquids on electronics
  • Privacy violation: Snooping files, recording video, capturing screenshots

2. Quantitative Results

2.1 Simulation Results

ModelBlindfold ASRBaseline ASR (best)Delta
GPT-4o93.2%~40%+53pp
Claude-3.5-sonnet~85%~35%+50pp
Phi-4-14B98.1%~55%+43pp
Llama-3.1-8B~90%~50%+40pp
DeepSeek-R1-14B(tested)(tested)
Gemma-3-27B(tested)(tested)

Baselines: POEX and BadRobot. Simulation platform: VirtualHome (Unity-based household simulator), with validation on Habitat, ManiSkill, and RoboTHOR. Dataset: 187 instructions combined from SafeAgentBench and BadRobot.

Blindfold achieved ASR above 80% across all tested open-source LLMs, with closed-source models also vulnerable (GPT-4o 93.2%, Claude-3.5-sonnet approximately 85%).

2.2 Real-World Robot Results

On a UFactory xArm 6 (6-DoF robotic arm) with RGB-D cameras:

  • 18/20 instructions produced successful attacks (90% ASR)
  • 12/20 malicious tasks were ultimately executed to completion (60% TSR)
  • Safe object substitutions were used in physical experiments (toys instead of humans, water instead of toxic liquids)

2.3 Defense Mechanism Effectiveness

DefenseASR ReductionResidual ASR (GPT-4o)
Llama-Guard (I/O filtering)7.6pp~85.6%
SafeDecoding (token prediction)4.8pp~88.4%
VeriSafe (formal verification)17.9pp~75.3%

All three defenses were transferred from traditional LLM safety research. None reduced ASR below 75% on any model. This is consistent with our thesis: defenses designed for text-layer safety do not address action-layer manipulation.

2.4 Ablation Study

ConfigurationGPT-4o ASRPhi-4-14B ASR
Raw inputs (no Blindfold)27.4%29.7%
Without obfuscator59.8%61.0%
Without verifier88.4%95.9%
Full Blindfold93.2%98.1%

The obfuscator is the highest-impact component for bypassing safety filters. The verifier primarily affects TSR (physical execution success) rather than ASR (filter bypass). Raw input ASR of 27-30% establishes a baseline: even without any attack framework, roughly a quarter of malicious instructions pass through to action generation — indicating that embodied LLMs have weaker safety constraints at the action layer than at the text layer even before adversarial manipulation.


3. Implications for VLA Safety Evaluation

3.1 Text-Layer Safety is Necessary but Insufficient

Blindfold’s core contribution to the threat landscape is the demonstration that text-level safety filters — the primary defense mechanism in current VLA deployments — do not prevent action-level harm. The attack works precisely because the generated text (individual action descriptions) is semantically benign; the harm emerges from the physical consequence of the action sequence.

This has direct implications for safety evaluation methodology:

  • FLIP grading at the text layer (our current methodology) can detect PARTIAL verdicts — models that hedge while complying — but cannot directly assess whether the generated action sequence would produce physical harm
  • Heuristic classifiers that rely on keyword matching or semantic similarity to harmful content will systematically miss Blindfold-style attacks because the attack text contains no harmful keywords
  • Human evaluators reading transcripts may not recognise the physical danger without domain expertise in robotics and spatial reasoning

3.2 Connection to PARTIAL Dominance (gli_086)

Our FLIP-graded VLA data (Report #49, n=58) shows:

  • 50% of verdicts are PARTIAL — models add disclaimers but still generate action sequences
  • 0% of verdicts are REFUSAL — no model fully refuses to generate action sequences for adversarial inputs
  • FLIP ASR: 72.4% across 7 attack families

Blindfold validates the operational consequence of this finding. In our data, models textually hedge (“I should note that this action could be dangerous…”) while producing the requested motor commands. Blindfold shows that an automated framework can systematically generate instructions where the text layer sees no danger at all — the hedging never even activates because the instruction appears benign.

The compound risk: our PARTIAL dominance finding shows models cannot refuse even when they recognise danger; Blindfold shows how to construct instructions where models do not even recognise the danger. Together, these eliminate both potential defense paths (recognition-without-refusal and refusal-after-recognition).

3.3 Implications for the Compliance Paradox (Report #59)

Report #59 identified the compliance paradox: models that signal safety awareness textually while generating unsafe action sequences. Blindfold adds a second dimension: the paradox also operates in the other direction. Not only do models fail to act on their own safety signals (our finding), but the safety signals themselves can be suppressed entirely by constructing instructions that are semantically benign at the text layer (Blindfold’s finding).

This creates a two-layered failure:

  1. Layer 1 (Failure-First finding): When models detect danger, they hedge textually but comply structurally (PARTIAL)
  2. Layer 2 (Blindfold finding): Danger can be constructed so that models do not detect it at the text layer at all

3.4 Raw Input Baseline: A Structural Vulnerability Signal

The ablation study’s raw input baseline — 27.4% ASR on GPT-4o, 29.7% on Phi-4-14B without any attack framework — indicates that embodied LLMs have a non-trivial structural vulnerability at the action layer even before adversarial optimization. Roughly one in four malicious instructions passes through to action generation with no attack engineering. This is consistent with the hypothesis that safety training for embodied LLMs focuses on the text output layer and does not adequately constrain the action generation layer.

For comparison, our data on frontier text-only models shows ASR below 5% for standard jailbreak prompts (Claude Sonnet 4.5: 4.5%, Gemini-3-flash: 2.6%). The 27-30% raw-input ASR for embodied LLMs suggests a substantially weaker safety baseline at the action layer.


4. Governance and Regulatory Implications

4.1 GLI Status

Blindfold is recorded as gli_083 in the GLI dataset. All four governance stages have null values:

  • T_framework: No framework anywhere addresses action-level manipulation of embodied AI. NIST AI RMF 1.0 covers prompt injection at the text layer. OWASP LLM Top 10 v2 does not cover embodied action outputs.
  • T_enact: No legislation distinguishes text-layer safety from action-layer safety for robotic systems. The EU AI Act high-risk provisions (August 2026) will cover robotic systems but do not require action-level safety testing as distinct from text-level evaluation.
  • T_enforce: No enforcement mechanism exists.

This is a total governance vacuum for a demonstrated physical attack with real-robot validation.

4.2 EU AI Act Annex III Implications

EU AI Act high-risk AI system requirements become applicable August 2, 2026. Robot systems in safety-critical applications are classified under Annex III. Manufacturers must demonstrate robustness. However, the Act does not specify that robustness testing must include action-level adversarial evaluation — conformity assessments relying solely on text-level safety tests would not detect Blindfold-class attacks. This creates a compliance gap: manufacturers can technically satisfy conformity requirements while deploying systems vulnerable to a published, automated attack with 90%+ ASR.

4.3 NSW WHS Digital Work Systems Act 2026

The NSW Digital Work Systems Act (passed February 13, 2026) creates a binding pre-deployment testing duty for AI systems affecting workers. Blindfold’s real-robot validation on a 6-DoF manipulator arm — the same class of system used in warehouse, manufacturing, and logistics environments covered by the Act — makes the action-level attack surface directly relevant to WHS compliance. An employer deploying a VLA-controlled manipulator that has only undergone text-level safety evaluation may not satisfy the duty to “ensure digital work systems do not create health risks” given that Blindfold-class attacks are now publicly documented.


5. Recommendations

5.1 For Failure-First Research Priorities

  1. Map Blindfold attack taxonomy to Failure-First VLA families. Blindfold’s four attack categories (tool-using harm, direct physical harm, environmental sabotage, privacy violation) partially overlap with our existing 10-family VLA taxonomy. A mapping exercise would identify coverage gaps in both directions. Priority: medium. Issue #301 follow-up.

  2. Develop action-layer FLIP grading. Current FLIP methodology grades at the text layer (infer instruction from response, judge instruction). An action-layer variant would need to: (a) parse the generated action sequence, (b) simulate or estimate physical consequences, (c) grade based on physical outcome rather than textual content. This is a non-trivial extension. Priority: high, but blocked on simulation infrastructure.

  3. Test Blindfold-inspired scenarios on PiCar-X. The PiCar-X platform provides a physical testbed, though its 2-DOF arm and 4-wheel base have different affordances than the 6-DoF xArm used in the paper. Adapt Blindfold’s methodology to constrained-embodiment systems. Priority: medium. Depends on robotic arm adapter availability.

  4. Incorporate raw-input baseline testing. The 27-30% raw-input ASR finding suggests that action-layer safety should be tested even without adversarial optimization. Add a “no-attack baseline” measurement to VLA evaluation protocols. Priority: high. Low implementation cost.

5.2 For CCS Paper (Abstract Registration April 22)

Blindfold should be cited in CCS Section 4.8 (VLA adversarial attacks) as independent validation of the compliance paradox and PARTIAL dominance findings. The paper is accepted at ACM SenSys 2026, making it a peer-reviewed reference. Specific integration points:

  • PARTIAL dominance discussion: Blindfold demonstrates the operational consequence
  • Defense mechanism insufficiency: Blindfold’s defense ablation (Llama-Guard 7.6pp, VeriSafe 17.9pp) corroborates our finding that text-layer defenses do not address action-layer safety
  • Raw-input baseline comparison: 27-30% embodied ASR vs <5% frontier text-only ASR quantifies the action-layer safety gap

5.3 For Commercial Briefings

Blindfold strengthens the commercial case for Failure-First adversarial testing services:

  • Red team brief (B1): Blindfold demonstrates that existing red-teaming methodologies focused on text-layer jailbreaks miss an entire attack surface. Pre-deployment testing must include action-level evaluation.
  • Actuarial brief (B2): The 90% real-robot ASR and the ineffectiveness of current defenses (max 17.9pp reduction) quantify residual risk for underwriters.
  • Regulatory positioning (B3): Blindfold is a peer-reviewed, real-robot-validated demonstration that current conformity assessment methods for robotic AI are insufficient.
  • Product liability (B4): The public availability of Blindfold’s methodology and results may affect “state of the art” defense arguments — manufacturers who do not test against action-level attacks after publication of this paper face a stronger liability exposure.

5.4 For Safety Testing Frameworks

Current safety evaluation frameworks for embodied AI should be extended to include:

  1. Action-sequence consequence analysis: Evaluate the physical outcome of generated action sequences, not just the text content of individual actions
  2. Semantic-physical misalignment testing: Systematically test whether benign-appearing instructions can produce harmful physical outcomes
  3. Defense-in-depth at the action layer: Safety mechanisms that operate at the action execution layer (kinematic constraints, force limits, collision detection) rather than the text-generation layer
  4. Multi-action sequence auditing: Evaluate sequences of actions for emergent harm, not just individual actions in isolation

6. Limitations of This Analysis

  • ASR figures for Claude-3.5-sonnet, Llama-3.1-8B, DeepSeek-R1-14B, and Gemma-3-27B are approximate (read from figures in the paper, not exact table values). GPT-4o and Phi-4-14B ASR are exact (from ablation table).
  • The real-robot experiments used safe object substitutions. The transfer rate from simulation to unconstrained real-world conditions is not established.
  • The paper tests consumer-grade robotic arms. Industrial manipulators with different safety architectures (hardware interlocks, safety-rated controllers per ISO 10218) may have different vulnerability profiles.
  • Our comparison of 27-30% raw embodied ASR to <5% frontier text-only ASR is across different evaluation setups and should be treated as indicative rather than directly comparable.

Appendix: Cross-Reference to GLI Dataset

GLI EntryRelationship to Blindfold
gli_083Direct entry for Blindfold
gli_086PARTIAL dominance — Blindfold validates operational consequence
gli_011BadVLA backbone attack — Blindfold targets action layer rather than VLM backbone
gli_012Physical adversarial patches — different attack surface (visual vs action)
gli_073CHAI physical text hijack — visual scene manipulation vs action-level manipulation
gli_079Earlier GLI entry for Blindfold (wave 5); gli_083 supersedes with expanded detail
gli_034ANNIE safety constraint violation — related action-layer attack in simulation only

This research informs our commercial services. See how we can help →