| Literature DB >> 33092893 |
Adam S Lowet1, Qiao Zheng2, Sara Matias1, Jan Drugowitsch3, Naoshige Uchida4.
Abstract
Learning about rewards and punishments is critical for survival. Classical studies have demonstrated an impressive correspondence between the firing of dopamine neurons in the mammalian midbrain and the reward prediction errors of reinforcement learning algorithms, which express the difference between actual reward and predicted mean reward. However, it may be advantageous to learn not only the mean but also the complete distribution of potential rewards. Recent advances in machine learning have revealed a biologically plausible set of algorithms for reconstructing this reward distribution from experience. Here, we review the mathematical foundations of these algorithms as well as initial evidence for their neurobiological implementation. We conclude by highlighting outstanding questions regarding the circuit computation and behavioral readout of these distributional codes.Entities:
Keywords: artificial intelligence; deep neural networks; dopamine; machine learning; population coding; reward
Year: 2020 PMID: 33092893 PMCID: PMC8073212 DOI: 10.1016/j.tins.2020.09.004
Source DB: PubMed Journal: Trends Neurosci ISSN: 0166-2236 Impact factor: 13.837