Socially Compliant Mobile Robot Navigation via Inverse Reinforcement Learning

Abstract
Mobile robots are increasingly populating our human environments. To interact with humans in a socially compliant way, these robots need to understand and comply with mutually accepted rules. In this paper, we present a novel approach to model the cooperative navigation behavior of humans. We model their behavior in terms of a mixture distribution that captures both the discrete navigation decisions, such as going left or going right, as well as the natural variance of human trajectories. Our approach learns the model parameters of this distribution that match, in expectation, the observed behavior in terms of user-defined features. To compute the feature expectations over the resulting high-dimensional continuous distributions, we use Hamiltonian Markov chain Monte Carlo sampling. Furthermore, we rely on a Voronoi graph of the environment to efficiently explore the space of trajectories from the robot’s current position to its target position. Using the proposed model, our method is able to imitate the behavior of pedestrians or, alternatively, to replicate a specific behavior that was taught by tele-operation in the target environment of the robot. We implemented our approach on a real mobile robot and demonstrated that it is able to successfully navigate in an office environment in the presence of humans. An extensive set of experiments suggests that our technique outperforms state-of-the-art methods to model the behavior of pedestrians, which also makes it applicable to fields such as behavioral science or computer graphics.

@article{kretzschmar16ijrr,
  author = {Henrik Kretzschmar and Markus Spies and Christoph Sprunk and Wolfram Burgard},
  title = {Socially Compliant Mobile Robot Navigation via Inverse Reinforcement Learning},
  journal = {The International Journal of Robotics Research},
  year = 2016,
  doi = {10.1177/0278364915619772}
}
Powered by bibtexbrowser
Back to Publications