Interactive Robot Learning of Gestures, Language and Affordances

Giovanni Saponaro, Lorenzo Jamone, Alexandre Bernardino, Giampiero Salvi

A growing field in robotics and Artificial Intelligence (AI) research is human–robot collaboration, whose target is to enable effective teamwork between humans and robots. However, in many situations human teams are still superior to human–robot teams, primarily because human teams can easily agree on a common goal with language, and the individual members observe each other effectively, leveraging their shared motor repertoire and sensorimotor resources. This paper shows that for cognitive robots it is possible, and indeed fruitful, to combine knowledge acquired from interacting with elements of the environment (affordance exploration) with the probabilistic observation of another agent’s actions. We propose a model that unites (i) learning robot affordances and word descriptions with (ii) statistical recognition of human gestures with vision sensors. We discuss theoretical motivations, possible implementations, and we show initial results which highlight that, after having acquired knowledge of its surrounding environment, a humanoid robot can generalize this knowledge to the case when it observes another agent (human partner) performing the same motor actions previously executed during training.

 DOI: 10.21437/GLU.2017-17

Cite as: Saponaro, G., Jamone, L., Bernardino, A., Salvi, G. (2017) Interactive Robot Learning of Gestures, Language and Affordances. Proc. GLU 2017 International Workshop on Grounding Language Understanding, 83-87, DOI: 10.21437/GLU.2017-17.

  author={Giovanni Saponaro and Lorenzo Jamone and Alexandre Bernardino and Giampiero Salvi},
  title={Interactive Robot Learning of Gestures, Language and Affordances},
  booktitle={Proc. GLU 2017 International Workshop on Grounding Language Understanding},