Multi-modal visual attention for robotics active vision systems - A reference architecture

Martin Hülse, Sebastian McBride, Mark Lee

Research output: Contribution to conferencePaper

1 Citation (Scopus)

Abstract

This work introduces an architecture for a robotic active vision system equipped with a manipulator that is able to integrate visual and non-visual (tactile) sensorimotor experiences. Inspired by the human vision system, we have implemented a strict separation of object location (where-data) and object features (what-data) in the visual data stream. This separation of what- and where-data has computational advantage but requires sequential fixation of visual cues in order to create and update a coherent view of the world. Hence, visual attention mechanisms must be put in place to decide which is the most task-relevant cue to fixate next. Regarding object manipulation many task relevant object properties (e.g. tactile feedback) are not necessarily related to visual features. Therefore, it is important that non-visual object features can influence visual attention. We present and demonstrate visual attention mechanisms for an active vision system that are modulated by visual and non-visual object features.
Original languageEnglish
Pages21-29
Number of pages9
Publication statusPublished - Apr 2011
EventProceedings of the AISB Symposium on Architectures for Active Vision - York, United Kingdom of Great Britain and Northern Ireland
Duration: 04 Apr 201107 Apr 2011

Conference

ConferenceProceedings of the AISB Symposium on Architectures for Active Vision
Country/TerritoryUnited Kingdom of Great Britain and Northern Ireland
CityYork
Period04 Apr 201107 Apr 2011

Fingerprint

Dive into the research topics of 'Multi-modal visual attention for robotics active vision systems - A reference architecture'. Together they form a unique fingerprint.

Cite this