| Literature DB >> 34026948 |
Ádám Koblinger1, József Fiser1, Máté Lengyel1,2.
Abstract
Perception is often described as probabilistic inference requiring an internal representation of uncertainty. However, it is unknown whether uncertainty is represented in a task-dependent manner, solely at the level of decisions, or in a fully Bayesian manner, across the entire perceptual pathway. To address this question, we first codify and evaluate the possible strategies the brain might use to represent uncertainty, and highlight the normative advantages of fully Bayesian representations. In such representations, uncertainty information is explicitly represented at all stages of processing, including early sensory areas, allowing for flexible and efficient computations in a wide variety of situations. Next, we critically review neural and behavioral evidence about the representation of uncertainty in the brain agreeing with fully Bayesian representations. We argue that sufficient behavioral evidence for fully Bayesian representations is lacking and suggest experimental approaches for demonstrating the existence of multivariate posterior distributions along the perceptual pathway.Entities:
Year: 2021 PMID: 34026948 PMCID: PMC8121756 DOI: 10.1016/j.cobeha.2021.03.009
Source DB: PubMed Journal: Curr Opin Behav Sci ISSN: 2352-1546
Figure 1Taxonomy of generative and recognition models in decision making. Blue and green backgrounds correspond to the two components of Bayesian decision theory: the observation and decision processes of the generative model (orange), and the perception and action selection modules of the recognition model (purple), respectively. Note that ‘perception’ here is broadly construed to include all cognitive processes (e.g. sensory perception or memory) that have access to information (‘observations’) that is relevant for the decision making task. Rectangles indicate observed variables (), circles indicate latent variables (including the decision variable, ) which are part of the generative model and are probabilistically computed in a recognition model, diamonds indicate non-probabilistically computed internal variables of a recognition models, hexagons indicate variables specific to the decision process: the action () and the utility obtained (). The utility function () is shown without a bounding box to indicate that it is a parameter that is constant across trials or time steps, while other quantities change over time or trials. Left: generative models. All generative models describe how is related to , and how it (exclusively) determines the obtained for a given (as parameterized by ). Simple generative models only have a single latent variable, . Complex generative models have multiple latent variables beside . Right: recognition models. All recognition models compute action from observations . Probabilistic recognition models compute a posterior over given (Equation 1), which they then combine with to compute (Equations 2a and 2b). The non-probabilistic recognition model computes directly from , without computing a posterior over , and without explicitly representing . Probabilistic recognition models are further subdivided based on what other variables are computed probabilistically while computing the posterior over : simple models do not have any other variables, complex models do, with fully Bayesian, hybrid, and task-dependent models probabilistically computing all variables, a subset of them, or none of them, respectively.
Compatibility of specific implementations of probabilistic recognition models with behavioral and neural data
| recognition model | implementation | behavioral data | neural data |
|---|---|---|---|
| task-dependent | DDM | psychometric and chronometric curves of perceptual decisions [ | decision-related ramping activity of LIP single cells [ |
| PPC | psychometric and chronometric curves of perceptual decisions [ | Poisson-like variability of cortical neurons [ | |
| DNN | object recognition and categorization performance [ | feature selectivity along the hierarchy of visual cortex [ | |
| hybrid | CRP | human categorization [ | – |
| fully Bayesian | belief propagation | bistable perception [ | tight balance between excitation and inhibition [ |
| extended PPC | – | anatomy and physiology of the olfactory bulb [ | |
| DDC | – | dopaminergic or hippocampal activity [ | |
| sampling | cue combination [ | various static [ | |
Details of the specific models are discussed in the main text. As we also note there, implementations appropriate for fully Bayesian recognition models could also implement task-dependent recognition models (but not vice versa). Thus, data listed here as supporting such implementations (e.g. multistable perception for sampling) does not necessarily provide support for fully Bayesian recognition models per se. Experiments providing support for fully Bayesian recognition models are discussed later (see also Table 2).
Behavioral evidence for probabilistic and fully Bayesian recognition models
| normative advantages | ||||
|---|---|---|---|---|
| recognition model | task-flexibility | information fusion | active sensing | learning |
| probabilistic versus non-probabilistic | [ | [ | [ | [74•75] |
| fully Bayesian versus task-dependent | [ | – | – | – |
Studies providing behavioral support that the brain's recognition model is probabilistic (top row) or fully-Bayesian (bottom row) are organized according to the normative advantage of the probabilistic or fully Bayesian recognition model they verified behaviorally (columns). Below we summarize the evidence for probabilistic recognition models only, the existing evidence (or lack thereof) for fully-Bayesian recognition models is discussed in the main text. Task-flexibility. Humans (or monkeys) were found to exhibit a high degree of task-flexibility in tasks requiring generalization either to new utility functions [69] or to new stimuli giving rise to posterior distributions that are qualitatively different from those previously experienced in the task [70]. Information fusion. Studies of information fusion showed that humans can near optimally combine two sources of information. Notably, in several classical cue-combination studies, participants had plenty of everyday experience with combining the information from the two tested sensory modalities to enhance the absolute accuracy of their decisions [16,17]. Thus, these tasks required only modest generalization, and as such could be solved with non-probabilistic recognition models. Other studies showed optimal information fusion even when participants needed to combine two sources of information that they had never combined before [19], or a sequence of observations while the number and informativeness of observations was varied [71]. In these cases, having a separate (non-probabilistic) recognition model for each task condition would be infeasible. Moreover, humans provided reliable uncertainty reports about their own performance when the difficulty of trials was modulated by multiple task parameters (e.g. the number and contrast of items in a scene) [65], suggesting that uncertainty reports were based on a unified representation of uncertainty (i.e. a single posterior distribution) rather than heuristic estimates corresponding to the different task parameters. Furthermore, in line with a unified uncertainty representation, the reported uncertainties also reliably predicted stimulus-independent fluctuations in performance over and above those controlled by experimentally defined cues [72]; Á Koblinger et al., 2019, COSYNE, conference]. Active sensing. Eye movements are almost never rewarded directly as such, but typically depend on participants’ inferences about the currently viewed stimulus. Thus, they offer an ideal test bed for assessing behavioral signatures of probabilistic representations. For example, while performing the same visual search task in widely different lighting conditions, humans near-optimally adjusted their eye-movements to the changed lighting conditions, suggesting that they could efficiently generalize their eye-movement strategies across a wide range of posteriors [73]. In a visual pattern categorization task, eye movements were also shown to be optimized for information search [24]. Critically, this eye movement strategy correctly took into account the constantly evolving posterior distribution that a probabilistic recognition model computed over pattern category (the decision variable) based on (the growing set of) previous fixations in a trial (observations). Learning. As non-probabilistic recognition models provide no principled basis for unsupervised learning, appropriate stimulus reliability-dependent updating of a recognition model can be taken as a hallmark of the recognition model being probabilistic. Such optimal updating was reported in a perceptual discrimination task without feedback, in which human participants used their uncertainty estimates about the stimulus (the decision variable) to correctly update their estimate of the base rate of the stimulus (a parameter of the recognition model) [74]. Similarly, in an economic decision task [75], participants near-optimally adjusted the learning speed of a dynamically fluctuating reward rate (decision variable), despite a lack of direct feedback about reward rates.