Partial recurrence enables robust and efficient computation
Listed in
This article is not in any list yet, why not save it to one of your lists.Abstract
Neural circuits are sparse and bidirectional. Meaning that signals flow from early sensory areas to later regions and back. Yet, between connected areas there exist some but not all pathways. How does this structure, somewhere between feedforward and fully recurrent, shape circuit function? To address this question, we designed a new recurrent neural network model in which a set of weight matrices (i.e. pathways) can be combined to generate every network structure between feedforward and fully recurrent. We term these architectures partially recurrent neural networks (pRNNs). We trained over 25,000 pRNNs on a novel set of reinforcement learning tasks, designed to mimic multisensory navigation, and compared their performance across multiple functional metrics. Our findings reveal three key insights. First, many architectures match or exceed the performance of fully recurrent networks, despite using as few as one-quarter the number of parameters; demonstrating that partial recurrence enables energy efficient, yet performant solutions. Second, each pathway’s functional impact is both task and circuit dependent. For instance, feedback connections enhance robustness to noise in some, but not all contexts. Third, different pRNN architectures learn solutions with distinct input sensitivities and memory dynamics, and these computational traits help to explain their functional capabilities. Overall, our results demonstrate that partial recurrence enables robust and efficient computation - a finding that helps to explain why neural circuits are sparse and bidirectional, and how these principles could inform the design of artificial systems.
1 Key Points
-
We introduce a new neural network model in which a set of weight matrices (i.e. pathways) can be combined to generate every network structure between feedforward and fully recurrent.
-
Many of these partially recurrent (pRNN) architectures perform as well as, or even better than, fully recurrent networks, in terms of their task performance, sample efficiency and robustness to various perturbations.
-
Each pathway’s functional impact is both task and circuit dependent. For example, input to output skip connections boost learning speed in some, but not all tasks or architectures.
-
We propose a conceptual model, whereby: networks with different structures learn solutions with distinct computational traits, which shape their function.