Visually grounded human-robot interaction is recognizedto be an essential ingredient of socially intelligent robots, and theintegration of vision and language increasingly attracts attention ofresearchers in diverse fields. However, most systems lack the capabilityto adapt and expand themselves beyond the preprogrammed setof communicative behaviors. Their linguistic capabilities are still farfrom being satisfactory which make them unsuitable for real-worldapplications. In this paper we will present a system in which a roboticagent can learn a grounded language model by actively interactingwith a human user. The model is grounded in the sense that meaningof the words is linked to a concrete sensorimotor experience of theagent, and linguistic rules are automatically extracted from the interactiondata. The system has been tested on the NAO humanoid robotand it has been used to understand and generate appropriate naturallanguage descriptions of real objects. The system is also capable ofconducting a verbal interaction with a human partner in potentiallyambiguous situations.
|Numero di pagine||16|
|Rivista||INTERNATIONAL JOURNAL OF COMPUTATIONAL LINGUISTICS RESEARCH|
|Stato di pubblicazione||Published - 2010|