Opposite and complementary roles of the two calcium thresholds for inducing LTP and LTD in models of striatal projection neurons
Curation statements for this article:-
Curated by eLife
eLife Assessment
This computational study constitutes an extension to prior work on biophysical calcium-based synaptic plasticity rules with metaplasticity, investigating how single neurons can learn to perform non-linear pattern classification. This important work presents a significantly simpler solution to the studied problem with potentially broad applicability, there is however incomplete evidence to support the core conclusions.
This article has been Reviewed by the following groups
Discuss this preprint
Start a discussion What are Sciety discussions?Listed in
- Evaluated articles (eLife)
Abstract
Synaptic plasticity has been shown to occur when calcium, flowing into the synapse due to incoming stimuli, surpasses a threshold level. This threshold level is modifiable through a process called metaplasticity. Some neurons, such as the striatal projection neurons, use different sources of calcium as the signal for synaptic strengthening (long-term potentiation, LTP) or weakening (long-term depression, LTD), resulting in them having two thresholds for inducing plasticity. In this study, we show opposite and complementary roles of metaplasticity in these two thresholds for inducing LTP and LTD on learning how to solve the linear and nonlinear feature binding problem (FBP and NFBP). In short, metaplasticity in one threshold (e.g. LTD) allows synaptic plasticity of the opposite type (e.g. LTP) to be properly expressed. This happens because metaplasticity in the LTD threshold protects strengthened synapses from weakening, thus allowing them to persistently increase during learning (and encode learned patterns). Similarly, metaplasticity in the LTP threhsold prevents weakened synapses from strengthening, thus allowing them to persistently decrease. Metaplasticity in both thresholds is necessary when synapses are clustered and the neuron needs to rely on supralinear dendritic integration for learning.
Article activity feed
-
eLife Assessment
This computational study constitutes an extension to prior work on biophysical calcium-based synaptic plasticity rules with metaplasticity, investigating how single neurons can learn to perform non-linear pattern classification. This important work presents a significantly simpler solution to the studied problem with potentially broad applicability, there is however incomplete evidence to support the core conclusions.
-
Reviewer #1 (Public review):
Summary:
This computational modelling study addresses the important question of how neurons can learn non-linear functions using biologically realistic plasticity mechanisms. The study extends the previous related work on metaplasticity by Khodadadi et al. (2025), using the same detailed biophysical model and basic study design, while significantly simplifying the synaptic plasticity rule by removing non-linearities, reducing the number of free parameters, and limiting plasticity to only excitatory synapses. The rule itself is supervised by the presence or absence of a binary dopamine reward signal, and gated by separate calcium-sensitive thresholds for potentiation and depression. The author shows that, when paired with a strong form of dendritic non-linearity called a "plateau potential" and appropriate …
Reviewer #1 (Public review):
Summary:
This computational modelling study addresses the important question of how neurons can learn non-linear functions using biologically realistic plasticity mechanisms. The study extends the previous related work on metaplasticity by Khodadadi et al. (2025), using the same detailed biophysical model and basic study design, while significantly simplifying the synaptic plasticity rule by removing non-linearities, reducing the number of free parameters, and limiting plasticity to only excitatory synapses. The rule itself is supervised by the presence or absence of a binary dopamine reward signal, and gated by separate calcium-sensitive thresholds for potentiation and depression. The author shows that, when paired with a strong form of dendritic non-linearity called a "plateau potential" and appropriate pre-existing dendritic clustering of features, this simpler learning mechanism can solve a non-linear classification task similar to the classic XOR logic operator, with equal or better performance than the previous publication. The primary claims of this publication are that metaplasticity is required for learning non-linear feature classification, and that simultaneous dynamics in two separate thresholds (for potentiation and depression) are critical in this process. By systematically studying the properties of a biophysically plausible supervised learning rule, this paper adds interesting insights into the mechanics of learning complex computations in single neurons.
Strengths:
The simplified form of the learning rule makes it easier to understand and study than previous metaplasticity rules, and makes the conclusions more generalizable, while preserving biological realism. Since similar biophysical mechanisms and dynamics exist in many different cell types across the whole brain, the proposed rule could easily be integrated into a wide range of computational models specializing in brain regions beyond the striatum (which is the focus of this study), making it of broad interest to computational neuroscientists. The general approach of systematically fixing or modifying each variable while observing the effects and interactions with other variables is sound and brings great clarity to understanding the dynamic properties and mechanics of the proposed learning rule.
Weaknesses:
General notes
(1) The credibility of the main claims is mainly limited by the very narrow range of model parameters that was explored, including several seemingly arbitrary choices that were not adequately justified or explored.
(2) The choice to use a morphologically detailed biophysical model, rather than a simpler multi-compartment model, adds a great deal of complexity that further increases uncertainty as to whether the conclusions can generalize beyond the specific choices of model and morphology studied in this paper.
(3) The requirement for pre-existing synaptic clustering, while not implausible, greatly limits the flexibility of this rule to solve non-linear problems more generally.
(4) In order to claim that two thresholds are truly necessary, the author would have to show that other well-known rules with a single threshold (e.g., BCM) cannot solve this problem. No such direct head-to-head comparisons are made, raising the question of whether the same task could be achieved without having two separate plasticity thresholds.
Specific notes
(1) Regarding the limited hyperparameter search:
(a) On page 5, the author introduces the upper LTP threshold Theta_LTP. It is not clear why this upper threshold is necessary when the weights are already bounded by w_max. Since w_max is just another hyperparameter, why not set it to a lower value if the goal is to avoid excessively strong synapses? The values of w_max and Theta_LTP appear to have been chosen arbitrarily, but this question could be resolved by doing a proper hyperparameter search over w_max in the absence of an upper Theta_LTP.
(b) The author does not explore the effect of having separate learning rates for theta_LTP and theta_LTD, which could also improve learning performance in the NFBP. A more comprehensive exploration of these parameters would make the inclusion of theta_max (and the specific value chosen) a lot less arbitrary.
(c) Figure 4 Supplements 3-4: The author shows results for a hyperparameter search of the learning rule parameters, which is important to see. However, the parameter search is very limited: only 3 parameter values were tried, and there is no explanation or rationale for choosing these specific parameters. In particular, the metaplasticity learning rates do not even span one order of magnitude. If the author wants to claim that the learning rule is insensitive to this parameter, it should be explored over a much broader range of values (e.g., something like the range [0.1-10]).
(2) Regarding the similarity to BCM, the author would ideally directly implement the BCM learning rule in their model, but at the least the author could have shown whether a slight variant of their rule presented here can be effective: for example having a single (plastic, not fixed) Ca-dependent threshold that applies to both LTP and LTD, with a single learning rate parameter.
(3) This paper is extremely similar (and essentially an extension) to the work of Khodadadi et al. (2025). Yet this paper is not mentioned at all in the introduction, and the relation between these papers is not made clear until the discussion, leaving me initially puzzled as to what problems this paper addresses that have not already been extensively solved. The introduction could be reworked to make this connection clearer while pointing out the main differences in approach (e.g., the important distinction between "boosting" nonlinearities and plateau potentials).
(4) The introduction is missing some citations of other recent work that has addressed single-neuron non-linear computation and learning, such as Gidon et al (2020); Jones & Kording (2021).
(5) Figure 1: The figure prominently features mGluR next to the CaV channel, but there is no mention of mGluR in the introduction. The introduction should be updated to include this.
(6) Could the author explain why there is a non-monotonic increase/decrease in the [Ca]_L in Figure 2B_4? Perhaps my confusion comes from not understanding what a single line represents. Does each line represent the [Ca] in a single spine (and if so, which spine), or is each line an average of all the spines in a given stim condition?
(7) Row 124 (page 4): L-type Ca microdomains (in which ions don't diffuse and therefore don't interact with Ca_NMDA) is a critical assumption of this model. The references for this appear only in the discussion, so when reading this paper, I found myself a bit confused about why the same ion is treated as two completely independent variables with separate dynamics. Highlighting the assumption (with citations) a bit more clearly in the results section when describing the rule would help with understanding.
(8) Row 149 (page 5): The current formulation of the update rule is not actually multiplicative. The fact that the update is weight-dependent alone does not make it a multiplicative rule, and judging by equation (1) it appears to simply be an additive rule with a weight regularization term that guarantees weight bounds. For example, a similar weight-dependent update is also a core component of BTSP (Milstein et al. 2021; Galloni et al. 2025), which is another well-known *additive* rule. An actual multiplicative rule implies that the update itself is applied via a multiplication, i.e. w_new = w_old * delta_w
For an example of a genuinely multiplicative rule, see: Cornford et al. 2024, "Brain-like learning with exponentiated gradients"). Multiplicative rules have very different properties to additive rules, since larger weights tend to grow quickly while small weights shrink towards 0.
(9) Equation 1 (page 5): Shouldn't the depression term be written as: (w_min - w)? This term would be negative if w is larger than w_min, leading to LTD. As it is written now, a large w and small w_min would just cause further potentiation instead of depression.
(10) In the introduction, the teaching signal is described in binary terms (DA peak, or DA pause), but in Equation 1, it actually appears to take on 3 different values. Could the author clarify what the difference is between a "DA pause" and the "no DA" condition? The way I read it, pause = absence of DA = no DA
(11) Figure 3: In these experimental simulations, DA feedback comes in 400ms after the stimulus. The author could motivate this choice a bit better and explain the significance of this delay. Clearly, the equations have a delta_t term, but as far as the learning algorithm is concerned, it seems like learning would be more effective at delta_t=0. Is the choice of 400ms mainly motivated by experimental observations? On a related note, is it meaningful that the 200ms delta_t before the next stimulus is shorter than the 400ms pause from the first stimulus? Wouldn't the DA that arrives shortly before a stimulus also have an effect on the learning rule?
(12) Figure 4C: How is it possible that the theta_LTP value goes higher than the upper threshold (dashed line)? Equation 3 implies that it should always be lower.
(13) Row 429 (page 11): The statement that "without metaplasticity the NFBP cannot be solved" is overly general and not supported by the evidence presented. There exist many papers in which people solve similar non-linear feature learning problems with Hebbian or other bio-plausible rules that don't have metaplasticity. A more accurate statement that can be made here is that the specific rule presented in this paper requires metaplasticity.
(14) The methods section does not make any mention of publicly available code or a GitHub repository. The author should add a link to the code and put some effort into improving the documentation so that others can more easily assess the code and reproduce the simulations.
-
Reviewer #2 (Public review):
Summary:
The manuscript proposes interesting synaptic plasticity rules grounded in experimental data. Its main features are:
(1) plasticity depends on local calcium concentration driven by presynaptic activity and is independent of somatic action potentials,
(2) the rules incorporate metaplasticity, and (3) they demonstrate how a single neuron could address the feature-binding problem at the dendritic level.
The work extends a previous study (https://doi.org/10.7554/eLife.97274.2), to which the author also contributed.
The author models two calcium thresholds (LTP/LTD) from two different calcium sources (NMDA/VGCC), and these thresholds are flexible (metaplasticity rule, similar to BCM), which is claimed to be necessary for successful learning of both FBP and NFBP (linear and nonlinear feature binding …
Reviewer #2 (Public review):
Summary:
The manuscript proposes interesting synaptic plasticity rules grounded in experimental data. Its main features are:
(1) plasticity depends on local calcium concentration driven by presynaptic activity and is independent of somatic action potentials,
(2) the rules incorporate metaplasticity, and (3) they demonstrate how a single neuron could address the feature-binding problem at the dendritic level.
The work extends a previous study (https://doi.org/10.7554/eLife.97274.2), to which the author also contributed.
The author models two calcium thresholds (LTP/LTD) from two different calcium sources (NMDA/VGCC), and these thresholds are flexible (metaplasticity rule, similar to BCM), which is claimed to be necessary for successful learning of both FBP and NFBP (linear and nonlinear feature binding problem with 1 or 2 patterns). The role of each threshold seems to be opposite and complementary. One extra condition has been added: an upper threshold for LTP. This threshold serves to stop synaptic strengthening once synapses are strong enough to evoke a plateau. With that, synapses are not strengthened to the maximal value, avoiding strong supralinear integration for irrelevant patterns.
Strengths:
The current model implements not only local synaptic plasticity but also metaplasticity and solves the FBP at the dendrite level. Another strong aspect of the model is that metaplasticity in the LTD threshold protects strengthened synapses from weakening. In this way, as the author mentioned, metaplasticity is able to protect learned patterns from being forgotten or weakened and prevent irrelevant patterns from being stored. This is a nice modelling example of metaplasticity being helpful in preventing the catastrophic interference or forgetting (as has been explicitly discussed in a recent article https://doi.org/10.1016/j.tins.2022.06.002 ). The author might want to briefly mention or emphasize this aspect of the model, which might be interesting also for the AI community.
Weaknesses:
(1) What is novel in the current paper as compared to Khodadadi et al. eLife 2025? That is not completely clear and should be made clearer. Is it only a minor difference related to the fact that the new learning rule has metaplasticity in both calcium thresholds and is simpler? This seems to be just an incremental increase in knowledge/methods. Can the author defend his paper against this point from the „devil's advocate"? How is the conclusion of the author in the abstract that „metaplasticity in both thresholds is necessary" reconcilable with his previous publication (Khodadadi et al. eLife 2025), in which only metaplasticity in one threshold was successful in solving the nonlinear feature binding problem?
(2) As far as I can judge without testing the model, metaplasticity causes thresholds to monotonically increase during systematic pattern presentation, which stabilizes weights and allows pattern separation. Due to the closed-loop nature of the current implementation, where metaplasticity only happens if plasticity happens, this also effectively locks patterns in place. However, flexible learning is an essential mechanism for survival. Imagine a mutation event takes place and bananas suddenly become red and/or strawberries turn yellow. It seems that the current model would be unable to adapt to these new patterns even if rewards were to be shifted. While out of the scope of the study, due to its importance, I feel that pattern shifting/relearning should at least be briefly discussed. How could the model be improved to allow relearning?
-
-
-