RISE
Using Non-Expert Data to Robustify Imitation
Learning via Offline Reinforcement Learning
Kevin Huang*,
Rosario Scalise*,
Cleah Winston,
Yunchu Zhang,
Rohan Baijal,
Ayush Agrawal,
Markus Grotz,
Bryon Boots,
Abhishek Gupta
(*:equal contribution)
University of Washington
Benjamin Burchfiel,
Hongkai Dai,
Masha Itkina,
Paarth Shah
Toyota Research Institute
RISE: Robust Imitation by Stiching from Experts demonstrates the following benefits.
Abstract.
Imitation learning has proven highly effective for training robots to
perform complex tasks from expert human demonstrations. However, it remains
limited by reliance on high-quality, task-specific data, restricting adaptability to
the diverse range of real-world object configurations and scenarios. In contrast,
non-expert data—such as play data, suboptimal demonstrations, or partial task
completions—can offer broader coverage and lower collection costs, but conven-
tional imitation learning approaches fail to utilize this data effectively. To address
these challenges, we show that offline reinforcement learning can be used as a
tool to harness non-expert data to enhance the performance of imitation learning
policies. We show that while standard offline RL approaches can be ineffective
at actually leveraging non-expert data under sparse coverage, simple algorithmic
modifications can allow the utilization of this data without significant additional
assumptions. Our approach shows that broadening the support of the policy dis-
tribution in offline RL can allow offline RL augmented imitation algorithms to
solve tasks robustly, under sparse coverage. In manipulation tasks, these innova-
tions significantly increase the range of states where learned policies are success-
ful when non-expert data is incorporated. Moreover, we show that these methods
are able to leverage all collected data, including partial or suboptimal demonstra-
tions, to bolster task-directed policy performance, underscoring the importance of
methods for using non-expert data for scalable and robust robot learning.
We introduce Robust Imitation Learning by Stitching from Experts,
or RISE 🌅.
Real-world Demo.
Approach.
0. Non-expert Data.
Figure 1: Another.
1. Offline-RL.
Figure 2: Another.
2. Lipschitz Continuity.
Figure 3: Another.
Acknowledgements.
The authors would like to acknowledge the members of the Robot Learning Lab and the Washington Embodied Intelligence and Robotics Development Lab for helpful and informative discussions throughout the process of this research.
The authors would also like to thank Emma Romig for robot hardware help. This research was supported by funding from Toyota Research Institute, under the University 2.0 research program.
BibTeX
Insert bibtex here.
}