| Literature DB >> 31047778 |
Yinan Cao1, Christopher Summerfield2, Hame Park3, Bruno Lucio Giordano4, Christoph Kayser5.
Abstract
When combining information across different senses, humans need to flexibly select cues of a common origin while avoiding distraction from irrelevant inputs. The brain could solve this challenge using a hierarchical principle by deriving rapidly a fused sensory estimate for computational expediency and, later and if required, filtering out irrelevant signals based on the inferred sensory cause(s). Analyzing time- and source-resolved human magnetoencephalographic data, we unveil a systematic spatiotemporal cascade of the relevant computations, starting with early segregated unisensory representations, continuing with sensory fusion in parietal-temporal regions, and culminating as causal inference in the frontal lobe. Our results reconcile previous computational accounts of multisensory perception by showing that prefrontal cortex guides flexible integrative behavior based on candidate representations established in sensory and association cortices, thereby framing multisensory integration in the generalized context of adaptive behavior.Entities:
Keywords: MEG; causal inference; crossmodal; decision making; flexible behavior; magnetoencephalography; parietal cortex; representational similarity analysis; sensory fusion; structure inference; ventrolateral prefrontal cortex
Mesh:
Year: 2019 PMID: 31047778 DOI: 10.1016/j.neuron.2019.03.043
Source DB: PubMed Journal: Neuron ISSN: 0896-6273 Impact factor: 17.173