Literature DB >> 20129865

An object-based visual attention model for robotic applications.

Yuanlong Yu1, George K I Mann, Raymond G Gosine.   

Abstract

By extending integrated competition hypothesis, this paper presents an object-based visual attention model, which selects one object of interest using low-dimensional features, resulting that visual perception starts from a fast attentional selection procedure. The proposed attention model involves seven modules: learning of object representations stored in a long-term memory (LTM), preattentive processing, top-down biasing, bottom-up competition, mediation between top-down and bottom-up ways, generation of saliency maps, and perceptual completion processing. It works in two phases: learning phase and attending phase. In the learning phase, the corresponding object representation is trained statistically when one object is attended. A dual-coding object representation consisting of local and global codings is proposed. Intensity, color, and orientation features are used to build the local coding, and a contour feature is employed to constitute the global coding. In the attending phase, the model preattentively segments the visual field into discrete proto-objects using Gestalt rules at first. If a task-specific object is given, the model recalls the corresponding representation from LTM and deduces the task-relevant feature(s) to evaluate top-down biases. The mediation between automatic bottom-up competition and conscious top-down biasing is then performed to yield a location-based saliency map. By combination of location-based saliency within each proto-object, the proto-object-based saliency is evaluated. The most salient proto-object is selected for attention, and it is finally put into the perceptual completion processing module to yield a complete object region. This model has been applied into distinct tasks of robots: detection of task-specific stationary and moving objects. Experimental results under different conditions are shown to validate this model.

Entities:  

Mesh:

Year:  2010        PMID: 20129865     DOI: 10.1109/TSMCB.2009.2038895

Source DB:  PubMed          Journal:  IEEE Trans Syst Man Cybern B Cybern        ISSN: 1083-4419


  3 in total

1.  Feedback-Driven Sensory Mapping Adaptation for Robust Speech Activity Detection.

Authors:  Ashwin Bellur; Mounya Elhilali
Journal:  IEEE/ACM Trans Audio Speech Lang Process       Date:  2016-12-13

2.  Event-driven proto-object based saliency in 3D space to attract a robot's attention.

Authors:  Suman Ghosh; Giulia D'Angelo; Arren Glover; Massimiliano Iacono; Ernst Niebur; Chiara Bartolozzi
Journal:  Sci Rep       Date:  2022-05-10       Impact factor: 4.996

3.  Combining segmentation and attention: a new foveal attention model.

Authors:  Rebeca Marfil; Antonio J Palomino; Antonio Bandera
Journal:  Front Comput Neurosci       Date:  2014-08-14       Impact factor: 2.380

  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.