RISE 🌅
Using Non-Expert Data to Robustify Imitation Learning via Offline Reinforcement Learning
Kevin Huang*, Rosario Scalise*, Cleah Winston, Yunchu Zhang, Rohan Baijal, Ayush Agrawal, Markus Grotz, Bryon Boots, Abhishek Gupta          (*:equal contribution) University of Washington UW Logo Benjamin Burchfiel, Hongkai Dai, Masha Itkina, Paarth Shah Toyota Research Institute Toyota Logo

RISE: Robust Imitation by Stiching from Experts demonstrates the following benefits.

Abstract.
Imitation learning has proven highly effective for training robots to perform complex tasks from expert human demonstrations. However, it remains limited by reliance on high-quality, task-specific data, restricting adaptability to the diverse range of real-world object configurations and scenarios. In contrast, non-expert data—such as play data, suboptimal demonstrations, or partial task completions—can offer broader coverage and lower collection costs, but conven- tional imitation learning approaches fail to utilize this data effectively. To address these challenges, we show that offline reinforcement learning can be used as a tool to harness non-expert data to enhance the performance of imitation learning policies. We show that while standard offline RL approaches can be ineffective at actually leveraging non-expert data under sparse coverage, simple algorithmic modifications can allow the utilization of this data without significant additional assumptions. Our approach shows that broadening the support of the policy dis- tribution in offline RL can allow offline RL augmented imitation algorithms to solve tasks robustly, under sparse coverage. In manipulation tasks, these innova- tions significantly increase the range of states where learned policies are success- ful when non-expert data is incorporated. Moreover, we show that these methods are able to leverage all collected data, including partial or suboptimal demonstra- tions, to bolster task-directed policy performance, underscoring the importance of methods for using non-expert data for scalable and robust robot learning. We introduce Robust Imitation Learning by Stitching from Experts, or RISE 🌅.
Real-world Demo.
Approach.

Figures explaining.

0. Non-expert Data.
real-to-sim pipeline

Figure 1: Another.

1. Offline-RL.
Approach Overview

Figure 2: Another.

2. Lipschitz Continuity.
Training in Sim

Figure 3: Another.

Acknowledgements.

The authors would like to acknowledge the members of the Robot Learning Lab and the Washington Embodied Intelligence and Robotics Development Lab for helpful and informative discussions throughout the process of this research. The authors would also like to thank Emma Romig for robot hardware help. This research was supported by funding from Toyota Research Institute, under the University 2.0 research program.

BibTeX
Insert bibtex here.
        }