Abstract

The random-trial incremental (RTI) model of human associative learning proposes that learning due to a trial where the association is presented proceeds incrementally; but with a certain probability, constant across trials, no learning occurs due to a trial. Based on RTI, identifying a policy for sequencing presentation trials of different associations for maximizing overall learning can be accomplished via a Markov decision process. For both finite and infinite horizons and a quite general structure of costs and rewards, a policy that on each trial presents an association that leads to the maximum expected immediate net reward is optimal.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call