Abstract

We present three artificial-grammar experiments. The first used position constraints, and the second used sequential constraints. The third varied both the amount of training and the degree of sequential constraint. Increasing both the amount of training and the redundancy of the grammar benefited participants' ability to infer grammatical status; nevertheless, they were unable to describe the grammar. We applied a multitrace model of memory to the task. The model used a global measure of similarity to assess the grammatical status of the probe and captured performance both in our experiments and in three classic studies from the literature. The model shows that retrieval is sensitive to structure in memory, even when individual exemplars are encoded sparsely. The work ties an understanding of performance in the artificial-grammar task to the principles used to understand performance in episodic-memory tasks.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call