Temporal Logic Imitation: Learning Plan-Satisficing Motion Policies from Demonstrations

TitleTemporal Logic Imitation: Learning Plan-Satisficing Motion Policies from Demonstrations
Publication TypeConference Paper
Year of Publication2022
AuthorsWang, Y., N. Figueroa, S. Li, A. Shah, and J. Shah
Conference Name6th Annual Conference on Robot Learning, Auckland, New Zealand
Date Published12/2022
KeywordsCertifiable Imitation Learning, Dynamical Systems, Formal Methods
AbstractLearning from demonstration (LfD) has succeeded in tasks featuring a long time horizon. However, when the problem complexity also includes humanin-the-loop perturbations, state-of-the-art approaches do not guarantee the successful reproduction of a task. In this work, we identify the roots of this challenge as the failure of a learned continuous policy to satisfy the discrete plan implicit in the demonstration. By utilizing modes (rather than subgoals) as the discrete abstraction and motion policies with both mode invariance and goal reachability properties, we prove our learned continuous policy can simulate any discrete plan specified by a linear temporal logic (LTL) formula. Consequently, an imitator is robust to both task- and motion-level perturbations and guaranteed to achieve task success. Project page: https://yanweiw.github.io/tli/
URLhttps://arxiv.org/abs/2206.04632
DOI10.48550/arXiv.2206.04632