Literature DB >> 32563023

Contextual encoder-decoder network for visual saliency prediction.

Alexander Kroner1, Mario Senden2, Kurt Driessens3, Rainer Goebel4.   

Abstract

Predicting salient regions in natural images requires the detection of objects that are present in a scene. To develop robust representations for this challenging task, high-level visual features at multiple spatial scales must be extracted and augmented with contextual information. However, existing models aimed at explaining human fixation maps do not incorporate such a mechanism explicitly. Here we propose an approach based on a convolutional neural network pre-trained on a large-scale image classification task. The architecture forms an encoder-decoder structure and includes a module with multiple convolutional layers at different dilation rates to capture multi-scale features in parallel. Moreover, we combine the resulting representations with global scene information for accurately predicting visual saliency. Our model achieves competitive and consistent results across multiple evaluation metrics on two public saliency benchmarks and we demonstrate the effectiveness of the suggested approach on five datasets and selected examples. Compared to state of the art approaches, the network is based on a lightweight image classification backbone and hence presents a suitable choice for applications with limited computational resources, such as (virtual) robotic systems, to estimate human fixations across complex natural scenes. Our TensorFlow implementation is openly available at https://github.com/alexanderkroner/saliency.
Copyright © 2020 The Author(s). Published by Elsevier Ltd.. All rights reserved.

Entities:  

Keywords:  Computer vision; Convolutional neural networks; Deep learning; Human fixations; Saliency prediction

Year:  2020        PMID: 32563023     DOI: 10.1016/j.neunet.2020.05.004

Source DB:  PubMed          Journal:  Neural Netw        ISSN: 0893-6080


  8 in total

1.  Hierarchical Multimodal Adaptive Fusion (HMAF) Network for Prediction of RGB-D Saliency.

Authors:  Ying Lv; Wujie Zhou
Journal:  Comput Intell Neurosci       Date:  2020-11-20

2.  Deep saliency models learn low-, mid-, and high-level features to predict scene attention.

Authors:  Taylor R Hayes; John M Henderson
Journal:  Sci Rep       Date:  2021-09-16       Impact factor: 4.379

3.  Linking Brain Structure, Activity, and Cognitive Function through Computation.

Authors:  Katrin Amunts; Javier DeFelipe; Cyriel Pennartz; Alain Destexhe; Michele Migliore; Philippe Ryvlin; Steve Furber; Alois Knoll; Lise Bitsch; Jan G Bjaalie; Yannis Ioannidis; Thomas Lippert; Maria V Sanchez-Vives; Rainer Goebel; Viktor Jirsa
Journal:  eNeuro       Date:  2022-03-11

4.  Semantic object-scene inconsistencies affect eye movements, but not in the way predicted by contextualized meaning maps.

Authors:  Marek A Pedziwiatr; Matthias Kümmerer; Thomas S A Wallis; Matthias Bethge; Christoph Teufel
Journal:  J Vis       Date:  2022-02-01       Impact factor: 2.240

5.  Construction of Home Product Design System Based on Self-Encoder Depth Neural Network.

Authors:  Guangpu Lu
Journal:  Comput Intell Neurosci       Date:  2022-04-21

6.  Object Categorization Capability of Psychological Potential Field in Perceptual Assessment Using Line-Drawing Images.

Authors:  Naoyuki Awano; Yuki Hayashi
Journal:  J Imaging       Date:  2022-03-26

7.  UNetGE: A U-Net-Based Software at Automatic Grain Extraction for Image Analysis of the Grain Size and Shape Characteristics.

Authors:  Ling Zeng; Tianbin Li; Xiekang Wang; Lei Chen; Peng Zeng; Jason Scott Herrin
Journal:  Sensors (Basel)       Date:  2022-07-26       Impact factor: 3.847

8.  A self-supervised deep neural network for image completion resembles early visual cortex fMRI activity patterns for occluded scenes.

Authors:  Michele Svanera; Andrew T Morgan; Lucy S Petro; Lars Muckli
Journal:  J Vis       Date:  2021-07-06       Impact factor: 2.240

  8 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.