Capacités audiovisuelles en robot humanoïde NAO

Jordi Sanchez-Riera 1
1 PERCEPTION - Interpretation and Modelling of Images and Videos
Inria Grenoble - Rhône-Alpes, LJK - Laboratoire Jean Kuntzmann, INPG - Institut National Polytechnique de Grenoble
Abstract : In this thesis we plan to investigate the complementarity of auditory and visual sensory data for building a high-level interpretation of a scene. The audiovisual (AV) input received by the robot is a function of both the external environment and of the robot's actual localization which is closely related to its actions. Current research in AV scene analysis has tended to focus on fixed perceivers. However, psychophysical evidence suggests that humans use small head and body movements, in order to optimize the location of their ears with respect to the source. Similarly, by walking or turning, the robot may be able to improve the incoming visual data. For example, in binocular perception, it is desirable to reduce the viewing distance to an object of interest. This allows the 3D structure of the object to be analyzed at a higher depth-resolution.
Document type :
Complete list of metadatas

Cited literature [92 references]  Display  Hide  Download
Contributor : Abes Star <>
Submitted on : Wednesday, March 12, 2014 - 9:27:53 AM
Last modification on : Thursday, February 7, 2019 - 1:23:07 AM
Long-term archiving on: Thursday, June 12, 2014 - 10:46:20 AM


Version validated by the jury (STAR)


  • HAL Id : tel-00953260, version 1



Jordi Sanchez-Riera. Capacités audiovisuelles en robot humanoïde NAO. Autre [cs.OH]. Université de Grenoble, 2013. Français. ⟨NNT : 2013GRENM009⟩. ⟨tel-00953260⟩



Record views


Files downloads