Project Website

Learning Safe-Stoppability Monitors for Humanoid Robots

Demo Video

Video of the learned safe-stop behavior, the monitor, and the loco-manipulation task used for the hardware and simulation studies.

Yifan Sun1 · Yiyuan Pan1 · Shangtao Li1 · Caiwu Ding2 · Tao Cui2 · Lingyun Wang2 · Changliu Liu1
1 Robotics Institute, Carnegie Mellon University
2 Foundational Technologies, Siemens Corporation

Humanoid emergency stops cannot simply cut power. PRISM learns a runtime monitor that predicts whether a fixed fallback controller can still drive the robot to a minimum-risk condition from the current state, and it refines that boundary efficiently with targeted importance sampling.

Policy-dependent reach-avoid safety Importance-sampled monitor refinement Simulation to real deployment
Carnegie Mellon University Robotics Institute logo
Intelligent Control Lab logo
Siemens logo

Overview

When is an emergency stop actually safe?

Instead of treating safety as an immediate shutdown, the paper studies whether a humanoid remains inside a safe-stoppable envelope: states from which a predefined fallback controller can still reach a safe terminal posture without falling or colliding.

"If we press the button, would the robot survive the stop?"

This is the core safe-stoppability question. The answer depends on the current state, the fixed fallback controller, and whether the robot can safely reach a minimum-risk condition.

Humanoid emergency stops are fundamentally different from power cutoffs used on fixed-base machines. A stop must preserve balance, manage contact, and steer the system toward a stable minimum-risk condition.

PRISM reframes this as a policy-dependent reach-avoid problem. Rather than certifying the entire state space, it focuses on the nominal task states the robot actually visits, then learns whether a safe stop remains feasible from each of those states.

Overview of safe-stoppability monitoring for a humanoid robot
Overview of the proposed pipeline. Safe-stop rollouts in a digital twin label whether the fallback controller can reach the minimum-risk condition from the current nominal state, and those labels are used to train a runtime safe-stoppability monitor.

Safe stop, not hard shutdown

The fallback controller is designed to preserve balance and drive the robot to a benign terminal posture instead of abruptly removing actuation.

Learn on nominal task states

The monitor is trained on the states encountered during task execution, which keeps the problem practical for high-dimensional humanoid systems.

Enable proactive intervention

The learned confidence can trigger the fallback policy before the robot crosses into states from which a safe stop is no longer possible.

Method

PRISM builds the monitor from simulation labels and targeted boundary search

The paper introduces PRISM, a simulation-driven framework for learning a neural safe-stoppability monitor under a fixed fallback controller. The key idea is to spend the simulation budget where failures are rare but most informative: near the safe versus unsafe boundary.

1

Collect nominal states

Record the state distribution induced by the task policy instead of attempting to cover the full humanoid state space.

2

Replay and label fallback outcomes

Reset the simulator to sampled states, execute the fallback controller, and label each state as safe or unsafe depending on whether the stop succeeds.

3

Train a neural monitor

Learn a state-conditioned predictor that estimates the probability that the safe stop will reach the terminal posture without intermediate failure.

4

Refine the boundary

Use importance sampling to focus new labels on uncertain or error-prone regions, improving unsafe-state recall under a fixed rollout budget.

PRISM overview showing data-efficient importance sampling near the stoppability boundary
PRISM reallocates dense sampling toward boundary regions where the current monitor is uncertain or wrong, rather than spending most of the labeling budget on trivially safe states.

Safe-Stoppability Monitor Learning

From replayed task states to real-time stop confidence

Each sampled state is evaluated by running the fallback controller forward until the robot either reaches the minimum-risk condition or violates safety constraints such as falls, forbidden collisions, or joint limits. Those outcomes become supervision for the monitor.

Unsafe stop failures are rare under nominal operation, which makes naive data collection inefficient. PRISM addresses this by iteratively retraining the monitor and then sampling more aggressively around the predicted safe versus unsafe boundary.

For sim-to-real transfer, nominal trigger states are taken from real robot logs and replayed in simulation before executing the fallback rollout. Domain randomization is then applied during fallback execution so the learned boundary remains useful on hardware.

SSE

Safe-stoppable envelope

The subset of nominal states from which the fallback controller can still reach the terminal safe set with high confidence.

MRC

Minimum-risk condition

A stable terminal posture used as the endpoint of the safe stop, chosen to minimize residual risk once the stop has been triggered.

Runtime use

Proactive intervention logic

The monitor can serve as a passive diagnostic after an external stop or as a proactive supervisor that triggers the fallback policy before irrecoverable states are reached.

Initial seed 3 trajectories

The iterative training loop starts from a small nominal dataset.

Per iteration +3 trajectories

PRISM grows the dataset gradually while reallocating where labels are collected.

Real logs 40 trajectories

Hardware nominal trajectories are replayed in simulation to reduce trigger-state mismatch.

Simulation seeds 400 seeds

Large-scale simulated rollouts provide the safe versus unsafe labels used for training.

Experiments / Results

Data efficiency, boundary quality, and sim-to-real consistency

Experiments evaluate the monitor on a 29-DoF humanoid performing a fixed loco-manipulation task. The paper emphasizes reducing false-safe predictions while keeping the total amount of hazardous data collection practical.

Unsafe accuracy 87.9%

At the same data volume as the 30-trajectory uniform baseline, PRISM improves unsafe-state prediction from 78.3% to 87.9%.

Data footprint >40% lower

PRISM reaches performance comparable to a 54-trajectory uniform baseline while using over 40% less collected data.

Real-world unsafe accuracy 94.1%

On held-out hardware trajectories, the importance-sampled monitor identifies unsafe states more reliably than the comparable uniform baseline.

Conservative tuning >99%

Threshold tuning can push simulated unsafe-state accuracy above 99% when more conservative intervention is preferred.

Experimental pipeline with real deployment, state reconstruction, and simulation rollouts
The experimental pipeline combines real humanoid execution, state reconstruction for replay, and large-scale fallback rollouts in simulation to label whether stopping succeeds from each sampled state.

Robust to many dynamics perturbations

The paper reports strong zero-shot generalization under joint damping and gain randomization. Ground friction changes are more difficult because they alter the contact mechanics that define recoverability.

Focused labels beat blanket dense sampling

Denser temporal sampling improves accuracy but becomes expensive quickly. PRISM is useful because it spends that extra density only where the current monitor still makes costly mistakes.

Paper

Paper and supporting media

The links on this page point to the latest manuscript PDF and demo video for this project.

Current Manuscript

Read the full paper

The arXiv PDF contains the full safe-stoppability formulation, the PRISM refinement procedure, the sim-to-real analysis, and the experimental tables summarized on this page.