A Fast Audiovisual Attention Model for Human Detection and Localization on a Companion Robot - AGPIG Accéder directement au contenu
Communication Dans Un Congrès Année : 2016

A Fast Audiovisual Attention Model for Human Detection and Localization on a Companion Robot

Résumé

This paper describes a fast audiovisual attention model applied to human detection and localization on a companion robot. Its originality lies in combining static and dynamic modalities over two analysis paths in order to guide the robot's gaze towards the most probable human beings' locations based on the concept of saliency. Visual, depth and audio data are acquired using a RGB-D camera and two horizontal microphones. Adapted state-of-the-art methods are used to extract relevant information and fuse them together via two dimensional gaussian representations. The obtained saliency map represents human positions as the most salient areas. Experiments have shown that the proposed model can provide a mean F-measure of 66 percent with a mean precision of 77 percent for human localization using bounding box areas on 10 manually annotated videos. The corresponding algorithm is able to process 70 frames per second on the robot.
Fichier principal
Vignette du fichier
16_Visual_ratajczak_.pdf (624.25 Ko) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)
Loading...

Dates et versions

hal-01408740 , version 1 (05-12-2016)

Licence

Copyright (Tous droits réservés)

Identifiants

  • HAL Id : hal-01408740 , version 1

Citer

Rémi Ratajczak, Denis Pellerin, Quentin Labourey, Catherine Garbay. A Fast Audiovisual Attention Model for Human Detection and Localization on a Companion Robot. VISUAL 2016 - The First International Conference on Applications and Systems of Visual Paradigms (VISUAL 2016), IARIA, Nov 2016, Barcelone, Spain. ⟨hal-01408740⟩
702 Consultations
336 Téléchargements

Partager

Gmail Facebook X LinkedIn More