TY - JOUR
T1 - Mapping language to vision in a real-world robotic scenario
AU - Stepanova, Karla
AU - Klein, Frederico B.
AU - Cangelosi, Angelo
AU - Vavrecka, Michal
N1 - Funding Information:
Manuscript received June 27, 2017; revised December 6, 2017; accepted February 24, 2018. Date of publication March 26, 2018; date of current version September 7, 2018. This work was supported in part by the European EU FP7 Research Project TRADR under Grant 609763, in part by TACR CAK under Grant TE01020197, in part by the CAPES Foundation, Ministry of Education of Brazil under Grant BEX 1084/13-5, in part by the CNPq Brazil under Grant 232590/2014-1, and in part by the U.K. EPSRC Project BABEL under Grant EP/J004561/1 and Grant EP/J00457X/1. (Corresponding author: Karla Štepánová.) K. Štepánová and M. Vavrecˇka are with the Czech Institute of Informatics, Robotics, and Cybernetics, Czech Technical University in Prague, 16000 Prague, Czech Republic (e-mail: [email protected]).
Publisher Copyright:
© 2016 IEEE.
PY - 2018/9/1
Y1 - 2018/9/1
N2 - Language has evolved over centuries and was gradually enriched and improved. The question, how people find assignment between meanings and referents, remains unanswered. There are many of computational models based on the statistical co-occurrence of meaning-reference pairs. Unfortunately, these mapping strategies show poor performance in an environment with a higher number of objects or noise. Therefore, we propose a more robust noise-resistant algorithm. We tested the performance of this novel algorithm with simulated and physical iCub robots. We developed a testing scenario consisting of objects with varying visual properties presented to the robot accompanied by utterances describing the given object. The results suggest that the proposed mapping procedure is robust, resistant against noise and shows better performance than one-step mapping for all levels of noise in the linguistic input, as well as slower performance degradation with increasing noise. Furthermore, the proposed procedure increases the clustering accuracy of both modalities.
AB - Language has evolved over centuries and was gradually enriched and improved. The question, how people find assignment between meanings and referents, remains unanswered. There are many of computational models based on the statistical co-occurrence of meaning-reference pairs. Unfortunately, these mapping strategies show poor performance in an environment with a higher number of objects or noise. Therefore, we propose a more robust noise-resistant algorithm. We tested the performance of this novel algorithm with simulated and physical iCub robots. We developed a testing scenario consisting of objects with varying visual properties presented to the robot accompanied by utterances describing the given object. The results suggest that the proposed mapping procedure is robust, resistant against noise and shows better performance than one-step mapping for all levels of noise in the linguistic input, as well as slower performance degradation with increasing noise. Furthermore, the proposed procedure increases the clustering accuracy of both modalities.
KW - cognitive modeling
KW - Computational modeling
KW - cross-situational learning
KW - Grounding
KW - iCub robot
KW - language acquisition.
KW - Linguistics
KW - Robots
KW - Robustness
KW - symbol grounding
KW - Task analysis
KW - Visualization
U2 - 10.1109/TCDS.2018.2819359
DO - 10.1109/TCDS.2018.2819359
M3 - Article
AN - SCOPUS:85044390731
SN - 2379-8920
VL - 10
SP - 784
EP - 794
JO - IEEE Transactions on Cognitive and Developmental Systems
JF - IEEE Transactions on Cognitive and Developmental Systems
IS - 3
M1 - 8325300
ER -