Abstract

Recurrent neural networks (RNNs) are powerful dynamical models, widely used in machine learning (ML) and neuroscience. Prior theoretical work has focused on RNNs with additive interactions. However gating i.e., multiplicative interactions are ubiquitous in real neurons and also the central feature of the best-performing RNNs in ML. Here, we show that gating offers flexible control of two salient features of the collective dynamics: (i) timescales and (ii) dimensionality. The gate controlling timescales leads to a novel marginally stable state, where the network functions as a flexible integrator. Unlike previous approaches, gating permits this important function without parameter fine-tuning or special symmetries. Gates also provide a flexible, context-dependent mechanism to reset the memory trace, thus complementing the memory function. The gate modulating the dimensionality can induce a novel, discontinuous chaotic transition, where inputs push a stable system to strong chaotic activity, in contrast to the typically stabilizing effect of inputs. At this transition, unlike additive RNNs, the proliferation of critical points (topological complexity) is decoupled from the appearance of chaotic dynamics (dynamical complexity). The rich dynamics are summarized in phase diagrams, thus providing a map for principled parameter initialization choices to ML practitioners.

Highlights

  • Recurrent neural networks (RNNs) are powerful dynamical systems that can represent a rich repertoire of trajectories and are popular models in neuroscience and machine learning

  • We introduce a gated RNN model that naturally extends a classical RNN by augmenting it with two kinds of gating interactions: (i) an update gate that acts like an adaptive time constant and (ii) an output gate which modulates the output of a neuron

  • We develop a theory for the gated RNN based on nonHermitian random matrix techniques [25,26] and the Martin–Siggia–Rose–De Dominicis–Janssen (MSRDJ) formalism [21,27–32] and use the theory to map out, in a phase diagram, the rich, functionally significant dynamical phenomena produced by gating

Read more

Summary

INTRODUCTION

Recurrent neural networks (RNNs) are powerful dynamical systems that can represent a rich repertoire of trajectories and are popular models in neuroscience and machine learning. We introduce a gated RNN model that naturally extends a classical RNN by augmenting it with two kinds of gating interactions: (i) an update gate that acts like an adaptive time constant and (ii) an output gate which modulates the output of a neuron The choice of these forms for gates are motivated by biophysical considerations The output gate allows fine control over the dimensionality of the network activity; control of the dimensionality can be useful during learning tasks [42] In certain regimes, this gate can mediate an input-driven chaotic transition, where static inputs can push a stable system abruptly to a chaotic state. Gates provide a flexible, contextdependent way to reset the state, providing a way to selectively erase the memory trace of past inputs We summarize these functionally significant phenomena in phase diagrams, which are practically useful for ML. Allow a principled and exhaustive exploration of dynamically distinct initializations

A RECURRENT NEURAL NETWORK MODEL TO STUDY GATING
HOW THE GATES SHAPE THE LINEARIZED DYNAMICS
Update gate facilitates slow modes and output gate causes instability
MARGINAL STABILITY AND ITS CONSEQUENCES
Condition for marginal stability
Functional consequences of marginal stability
A NOVEL CHAOTIC TRANSITION
Long-time behavior of the network
DMFT prediction for λmax
Condition for continuous transition to chaos
Output gate induces a novel chaotic transition
Spontaneous emergence of fixed points
Long chaotic transients
An input-induced chaotic transition
A FLEXIBLE RESET MECHANISM
PHASE DIAGRAMS FOR THE GATED NETWORK
Role of biases and static inputs
VIII. DISCUSSION
Significance of the update gate
Significance of the output gate
G G G ðA12Þ
Disorder averaging
XZ zðtÞ
Spontaneous emergence of fixed-points
Influence of update gate on the discontinuous transition
Effect of biases on the phase boundaries
Perturbative solutions for the fixed-point variance Δh with biases
ChðτÞ near critical point
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call