Skip to main content

Research Repository

Advanced Search

Learning Non-local Dependencies

Kuhn, G.; Dienes, Z.

Authors

G. Kuhn

Z. Dienes



Abstract

This paper addresses the nature of the temporary storage buffer used in implicit or statistical learning. Kuhn and Dienes [Kuhn, G., & Dienes, Z. (2005). Implicit learning of nonlocal musical rules: implicitly learning more than chunks. Journal of Experimental Psychology-Learning Memory and Cognition, 31(6) 1417–1432] showed that people could implicitly learn a musical rule that was solely based on non-local dependencies. These results seriously challenge models of implicit learning that assume knowledge merely takes the form of linking adjacent elements (chunking). We compare two models that use a buffer to allow learning of long distance dependencies, the Simple Recurrent Network (SRN) and the memory buffer model. We argue that these models – as models of the mind – should not be evaluated simply by fitting them to human data but by determining the characteristic behaviour of each model. Simulations showed for the first time that the SRN could rapidly learn non-local dependencies. However, the characteristic performance of the memory buffer model rather than SRN more closely matched how people came to like different musical structures. We conclude that the SRN is more powerful than previous demonstrations have shown, but it’s flexible learned buffer does not explain people’s implicit learning (at least, the affective learning of musical structures) as well as fixed memory buffer models do.

Citation

Kuhn, G., & Dienes, Z. (2008). Learning Non-local Dependencies. Cognition, 106(1), 184-206. https://doi.org/10.1016/j.cognition.2007.01.003

Journal Article Type Article
Online Publication Date Mar 6, 2007
Publication Date Jan 1, 2008
Deposit Date Feb 24, 2009
Journal Cognition
Print ISSN 0010-0277
Publisher Elsevier
Peer Reviewed Peer Reviewed
Volume 106
Issue 1
Pages 184-206
DOI https://doi.org/10.1016/j.cognition.2007.01.003
Keywords Implicit learning, Statistical learning, Artificial grammar learning, Chunks, Non-local dependencies, Simple recurrent network, Memory buffer model.