Home // VISUAL 2016, The First International Conference on Applications and Systems of Visual Paradigms // View article


A Fast Audiovisual Attention Model for Human Detection and Localization on a Companion Robot

Authors:
Rémi Ratajczak
Denis Pellerin
Catherine Garbay
Quentin Labourey

Keywords: audiovisual attention; saliency; RGB-D; human localization; companion robot

Abstract:
This paper describes a fast audiovisual attention model applied to human detection and localization on a companion robot. Its originality lies in combining static and dynamic modalities over two analysis paths in order to guide the robot’s gaze towards the most probable human beings’ locations based on the concept of saliency. Visual, depth and audio data are acquired using a RGB-D camera and two horizontal microphones. Adapted state-of-the-art methods are used to extract relevant information and fuse them together via two dimensional gaussian representations. The obtained saliency map represents human positions as the most salient areas. Experiments have shown that the proposed model can provide a mean F-measure of 66 percent with a mean precision of 77 percent for human localization using bounding box areas on 10 manually annotated videos. The corresponding algorithm is able to process 70 frames per second on the robot.

Pages: 30 to 35

Copyright: Copyright (c) IARIA, 2016

Publication date: November 13, 2016

Published in: conference

ISSN: 2519-8645

ISBN: 978-1-61208-520-3

Location: Barcelona, Spain

Dates: from November 13, 2016 to November 17, 2016