Unconscious reinforcement learning of hidden brain states supported by confidence

Aurelio Cortese, Hakwan Lau, Mitsuo Kawato

Research output: Contribution to journalArticlepeer-review

32 Scopus citations

Abstract

Can humans be trained to make strategic use of latent representations in their own brains? We investigate how human subjects can derive reward-maximizing choices from intrinsic high-dimensional information represented stochastically in neural activity. Reward contingencies are defined in real-time by fMRI multivoxel patterns; optimal action policies thereby depend on multidimensional brain activity taking place below the threshold of consciousness, by design. We find that subjects can solve the task within two hundred trials and errors, as their reinforcement learning processes interact with metacognitive functions (quantified as the meaningfulness of their decision confidence). Computational modelling and multivariate analyses identify a frontostriatal neural mechanism by which the brain may untangle the ‘curse of dimensionality’: synchronization of confidence representations in prefrontal cortex with reward prediction errors in basal ganglia support exploration of latent task representations. These results may provide an alternative starting point for future investigations into unconscious learning and functions of metacognition.

Original languageEnglish
Article number4429
JournalNature Communications
Volume11
Issue number1
DOIs
StatePublished - 1 Dec 2020
Externally publishedYes

Fingerprint

Dive into the research topics of 'Unconscious reinforcement learning of hidden brain states supported by confidence'. Together they form a unique fingerprint.

Cite this