Abstract

Within computational neuroscience, the algorithmic and neural basis of structure learning remains poorly understood. Concept learning is one primary example, which requires both a type of internal model expansion process (adding novel hidden states that explain new observations), and a model reduction process (merging different states into one underlying cause and thus reducing model complexity via meta-learning). Although various algorithmic models of concept learning have been proposed within machine learning and cognitive science, many are limited to various degrees by an inability to generalize, the need for very large amounts of training data, and/or insufficiently established biological plausibility. Using concept learning as an example case, we introduce a novel approach for modeling structure learning—and specifically state-space expansion and reduction—within the active inference framework and its accompanying neural process theory. Our aim is to demonstrate its potential to facilitate a novel line of active inference research in this area. The approach we lay out is based on the idea that a generative model can be equipped with extra (hidden state or cause) “slots” that can be engaged when an agent learns about novel concepts. This can be combined with a Bayesian model reduction process, in which any concept learning—associated with these slots—can be reset in favor of a simpler model with higher model evidence. We use simulations to illustrate this model's ability to add new concepts to its state space (with relatively few observations) and increase the granularity of the concepts it currently possesses. We also simulate the predicted neural basis of these processes. We further show that it can accomplish a simple form of “one-shot” generalization to new stimuli. Although deliberately simple, these simulation results highlight ways in which active inference could offer useful resources in developing neurocomputational models of structure learning. They provide a template for how future active inference research could apply this approach to real-world structure learning problems and assess the added utility it may offer.

Highlights

  • The ability to learn the latent structure of one’s environment— such as inferring the existence of hidden causes of regularly observed patterns in co-occurring feature observations—is central to human cognition

  • We provide a number of example simulations demonstrating how structure learning can be seen as an emergent property of active inference under generative models with “spare capacity”; where spare or uncommitted capacity is used to expand the repertoire of representations (Baker and Tenenbaum, 2014), while Bayesian model reduction (Hobson and Friston, 2012; Friston et al, 2017b) promotes generalization by minimizing model complexity—and releasing representations to replenish “spare capacity.”

  • Our goal is to present an introductory proof of concept— that could be used as the foundation of future active inference research on the neurocomputational basis of structure learning

Read more

Summary

INTRODUCTION

The ability to learn the latent structure of one’s environment— such as inferring the existence of hidden causes of regularly observed patterns in co-occurring feature observations—is central to human cognition. The structure of the space of hidden causes itself needs to expand to accommodate new patterns of observations This model expansion process is complementary to a process called Bayesian model reduction (Friston and Penny, 2011), in which the agent can infer that there is redundancy in its model, and a model with fewer states or parameters provides a more parsimonious (i.e., simpler) explanation of observations (Schmidhuber, 2006; Friston et al, 2017b). From a machine learning perspective, Bayesian model reduction affords the opportunity to consider generative models with a large number of hidden states or latent factors and optimize the number (or partitions) of latent factors with respect to a variational bound on model evidence This could be regarded as a bounded form of non-parametric Bayes, in which a potentially infinite number of latent factors are considered; with appropriate (e.g., Indian buffet process) priors over the number of hidden states generating data features. AN ACTIVE INFERENCE APPROACH FOR MODELING CONCEPT LEARNING THROUGH STATE-SPACE EXPANSION

A Primer on Active Inference
A Model of Concept Inference and Learning Through State-Space Expansion
Findings
DISCUSSION
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call