Jointly Improving Parsing and Perception for Natural Language Commands through Human-Robot Dialog

J. Thomason et al., “Jointly Improving Parsing and Perception for Natural Language Commands through Human-Robot Dialog,” The Journal of Artificial Intelligence Research (JAIR), vol. 67, Feb. 2020.

Abstract

In this work, we present methods for using human-robot dialog to improve language understanding for a mobile robot agent. The agent parses natural language to underlying semantic meanings and uses robotic sensors to create multi-modal models of perceptual concepts like red and heavy. The agent can be used for showing navigation routes, delivering objects to people, and relocating objects from one location to another. We use dialog clarification questions both to understand commands and to generate additional parsing training data. The agent employs opportunistic active learning to select questions about how words relate to objects, improving its understanding of perceptual concepts. We evaluated this agent on Amazon Mechanical Turk. After training on data induced from conversations, the agent reduced the number of dialog questions it asked while receiving higher usability ratings. Additionally, we demonstrated the agent on a robotic platform, where it learned new perceptual concepts on the fly while completing a real-world task.

BibTeX Entry

@article{thomason2020improving,
  title = {Jointly Improving Parsing and Perception for Natural Language Commands through Human-Robot Dialog},
  author = {Thomason, Jesse and Padmakumar, Aishwarya and Sinapov, Jivko and Walker, Nick and Jiang, Yuqian and Yedidsion, Harel and Hart, Justin and Stone, Peter and Mooney, Raymond J.},
  volume = {67},
  journal = {The Journal of Artificial Intelligence Research (JAIR)},
  month = feb,
  year = {2020},
  wwwtype = {journal},
  wwwpdf = {https://www.cs.utexas.edu/users/ml/papers/thomason.jair20.pdf}
}