Back to Search
Start Over
An algorithmic account for how humans efficiently learn, transfer, and compose hierarchically structured decision policies.
- Source :
-
Cognition [Cognition] 2025 Jan; Vol. 254, pp. 105967. Date of Electronic Publication: 2024 Oct 04. - Publication Year :
- 2025
-
Abstract
- Learning structures that effectively abstract decision policies is key to the flexibility of human intelligence. Previous work has shown that humans use hierarchically structured policies to efficiently navigate complex and dynamic environments. However, the computational processes that support the learning and construction of such policies remain insufficiently understood. To address this question, we tested 1026 human participants, who made over 1 million choices combined, in a decision-making task where they could learn, transfer, and recompose multiple sets of hierarchical policies. We propose a novel algorithmic account for the learning processes underlying observed human behavior. We show that humans rely on compressed policies over states in early learning, which gradually unfold into hierarchical representations via meta-learning and Bayesian inference. Our modeling evidence suggests that these hierarchical policies are structured in a temporally backward, rather than forward, fashion. Taken together, these algorithmic architectures characterize how the interplay between reinforcement learning, policy compression, meta-learning, and working memory supports structured decision-making and compositionality in a resource-rational way.<br /> (Copyright © 2024 The Authors. Published by Elsevier B.V. All rights reserved.)
Details
- Language :
- English
- ISSN :
- 1873-7838
- Volume :
- 254
- Database :
- MEDLINE
- Journal :
- Cognition
- Publication Type :
- Academic Journal
- Accession number :
- 39368350
- Full Text :
- https://doi.org/10.1016/j.cognition.2024.105967