Contrast-dependent response modulation in convolutional neural networks captures behavioral and neural signatures of visual adaptation

Read the full article See related articles

Listed in

This article is not in any list yet, why not save it to one of your lists.
Log in to save this article

Abstract

Human perception remains robust under challenging viewing conditions. Robust perception is thought to be facilitated by nonlinear response properties, including temporal adaptation (reduced responses to re-peated stimuli) and contrast gain (shift in the contrast response function with pre-exposure to a stimulus). Temporal adaptation and contrast gain have both been shown to aid object recognition, however, their joint effect on perceptual and neural responses remains unclear. Here, we collected behavioural measurements and electrocorticography (EEG) data while human participants (both sexes) classified objects embedded within temporally repeated noise patterns, whereby object contrast was varied. Our findings reveal an in-teraction effect, with increased categorization performance as a result of temporal adaptation for higher but not lower contrast stimuli. This increase in behavioral performance after adaptation is associated with more pronounced contrast-dependent modulation of evoked neural responses, as well as better decoding of object information from EEG activity. To elucidate the neural computations underlying these effects, we endowed deep convolutional neural networks (DCNN) with various temporal adaptation mechanisms, including intrinsic suppression and temporal divisive normalisation. We demonstrate that incorporating a biologically-inspired contrast response function to modify temporal adaptation helps DCNNs to accurately capture human behaviour and neural activation profiles. Moreover, we find that networks with multiplicative temporal adaptation mechanisms, such as divisive normalization, show higher robustness against spatial shifts in the inputs compared to DCNNs employing additive mechanisms. Overall, we reveal how interaction effects between nonlinear response properties influence human perception in challenging viewing contexts and investigate potential computations that mediate these effects.

Significance statement

Humans are able to perceive the environment even when viewing conditions are suboptimal. This robust perception has been linked to nonlinear neural processing of incoming visual information. Here, we examine the joint impact of two neural response properties, temporal adaptation and contrast gain, during object recognition, demonstrating interaction effects on categorization performance and in evoked neural responses. Using convolutional neural networks, we investigate various temporal adaptation mechanisms mediating the neural responses and perception, demonstrating that introducing contrast-dependent modulation of the unit activations captures human behaviour and neural object representations. Our findings shed light on how neural response properties give rise to robust perception and offer a framework to study the underlying neural dynamics and their impact on perception.

Article activity feed