RaC: Robot Learning for Long-Horizon Tasks by Scaling Recovery and Correction

1Carnegie Mellon University;
TL;DR: RaC achieves efficient performance scaling via a "mid-training" phase added after imitation pre-training. In this phase, RaC scales recovery & correction data via human intervention data collection. Learning on diverse skills data enables policies to reset, retry, and adapt, mitigating compounding errors in long-horizon manipulation tasks.

RaC standardizes human intervention data collection protocols with two rules:

  1. Upon intervention, operators first recover back to in-distribution states, then perform correction to complete the sub-task.
  2. After intervention completes, operator terminates the episode to prevent further rollouts.

On challenging real world long-horizon manipulation tasks, RaC achieves high success rates with efficient data scaling. It outperforms prior human-in-the-loop method HG-DAgger and conventional batched full expert trajectories data collection.

Performance comparison plot

Compared to previous SOTA results on similar shirt-hanging task, RaC outperforms prior approaches with ~10x less data.

shirt-hanging comparison table

During deployment, RaC exhibits test-time scaling characteristic — full task completion success rises with more recovery behaviors.

Test Time Scaling plot

RaC Data Collection Protocol

(Click Video to Play)

Recovery & Correction Improves Robustness to Failures

All Videos are autonomous policy rollouts recorded at 1x speed.

Learned Policies Comparison

(Click Video to Play)

Shirt Hanging

Airtight Container Lid Sealing

Clamshell Takeout Box Packing

Uncut Shirt Hanging Full Evaluation Recording (60 Trials, 2 Hours)

Out of 60 evaluation trials, the robot achieved 48 full task successes, 80% task success rate.

BibTeX

          
            @misc{hu2025racrobotlearninglonghorizon,
              title={RaC: Robot Learning for Long-Horizon Tasks by Scaling Recovery and Correction}, 
              author={Zheyuan Hu and Robyn Wu and Naveen Enock and Jasmine Li and Riya Kadakia and Zackory Erickson and Aviral Kumar},
              year={2025},
              eprint={2509.07953},
              archivePrefix={arXiv},
              primaryClass={cs.RO},
              url={https://arxiv.org/abs/2509.07953}, 
            }