Optimal Control of Spiking Neural Networks

Read the full article See related articles

Listed in

This article is not in any list yet, why not save it to one of your lists.
Log in to save this article

Abstract

Control theory provides a natural language to describe multi-areal interactions and flexible cognitive tasks such as covert attention or brain-machine interface (BMI) experiments, which require finding adequate inputs to a local circuit in order to steer its dynamics in a context-dependent manner. In optimal control, the target dynamics should maximize a notion of long-term value along trajectories, possibly subject to control costs. Because this problem is, in general, not tractable, current approaches to the control of networks mostly consider simplified settings (e.g., variations of the Linear-Quadratic Regulator). Here, we present a mathematical framework for optimal control of recurrent networks of stochastic spiking neurons with low-rank connectivity. An essential ingredient is a control-cost that penalizes deviations from the default dynamics of the network (specified by its recurrent connections), which motivates the controller to use the default dynamics as much as possible. We derive a Bellman Equation that specifies a Value function over the low-dimensional network state (LDS), and a corresponding optimal control input. The optimal control law takes the form of a feedback controller that provides external excitatory (inhibitory) synaptic input to neurons in the recurrent network if their spiking activity tends to move the LDS towards regions of higher (lower) Value. We use our theory to study the problem of steering the state of the network towards particular terminal regions which can lie either in or out of regions in the LDS with slow dynamics, in analogy to standard BMI experiments. Our results provide the foundation of a novel approach with broad applicability that unifies bottom-up and top-down perspectives on neural computation.

Article activity feed