The Training Village: an open platform for continuous testing of rodents in cognitive tasks

Curation statements for this article:
  • Curated by eLife

    eLife logo

    eLife Assessment

    This study introduces the "Training Village," a valuable system for which solid evidence shows that it enables group-housed rodents to autonomously learn complex tasks while preserving natural social interactions. The platform is flexible, allowing animals to learn multiple tasks sequentially and supporting applications in continual learning. This approach is likely to be of broad interest to behavioral researchers using rodent models in systems and cognitive neuroscience.

This article has been Reviewed by the following groups

Read the full article

Discuss this preprint

Start a discussion What are Sciety discussions?

Abstract

The development of new methods for dissecting neural circuits in rodents promises to transform the study of higher-order cognitive functions. However, the adaptation of complex cognitive tasks from humans to rodents remains very limited as these tasks require long periods of animal training with significant researcher expertise and effort. Multi-box setups can mitigate the problem but remain labor-intensive and costly to maintain. Newly developed home cage training systems, while minimizing costs, are limited to specific conditions and tasks. Here, we present the Training Village (TV), an open-source, affordable, and fully automated system for continuous rodent training. In the TV, a group of rodents lives in enriched arenas while individually accessing an operant box at any time. The system provides personalized training, continuous monitoring of task performance and home cage activity, and real-time remote supervision. Its graphical interface and modular design make it user-friendly and easy to integrate with other behavioral systems running a wide range of cognitive paradigms. We validated the TV across multiple tasks and cohorts of mice and rats, demonstrating efficient operant box usage, stable long-term task engagement, and its potential to link home cage behavior with task performance. Overall, the TV is an easy-to-adopt platform that can greatly accelerate brain research by fully automating rodent training in complex cognitive tasks.

Article activity feed

  1. eLife Assessment

    This study introduces the "Training Village," a valuable system for which solid evidence shows that it enables group-housed rodents to autonomously learn complex tasks while preserving natural social interactions. The platform is flexible, allowing animals to learn multiple tasks sequentially and supporting applications in continual learning. This approach is likely to be of broad interest to behavioral researchers using rodent models in systems and cognitive neuroscience.

  2. Reviewer #1 (Public review):

    Summary:

    The authors introduce the Training Village (TV), an open-source and modular system that allows group-housed rodents to live in enriched home cages while individually accessing a single shared operant box for automated cognitive training. The paper reported the animals' activity both in the operant box and in the home cages, which is novel.

    Strengths:

    A major strength of the work is that it moves beyond a proof-of-concept and demonstrates sustained box usage, long-term trial accumulation, and compatibility with different task designs.

    (1) The platform provided a technical contribution in rodent cognitive neuroscience: obtaining large amounts of behavioral data from complex tasks while reducing experimenter intervention and preserving social housing.

    (2) The authors demonstrate that the system can sustain prolonged task engagement (up to 12 months), maintain efficient use of a single operant box.

    (3) The manuscript opens interesting opportunities for studying behavior outside standard session-based training. Because animals self-initiate training while remaining in a group-housed setting, the platform has the potential to illuminate relationships among motivation, spontaneous activity, and task engagement that are hard to access in conventional paradigms.

    Weaknesses:

    (1) One area that would benefit from further clarification is the manuscript's core advance relative to prior automated group-housed training systems, particularly Mouse Academy (Qiao et al., 2018). The authors listed some advantages in the Discussion section; however, those were some minor engineering improvements, and what is more interesting is the scientific question or results that can be asked or obtained from this study. The current study clearly presents a functional and carefully documented platform, but it would help the reader if the authors more explicitly distinguished the present system from earlier related approaches, both in terms of system design and in terms of experimental validation.

    (2) At the system level, several of the claimed advantages could be supported more directly with quantitative data. For example, if the double-detection corridor and alarm system are important distinguishing features, it would be valuable to report measures such as detection accuracy, missed detections, co-entry failures, alarm frequency, and the degree of manual intervention required in practice. Similarly, the welfare-related arguments are plausible and important, but would be strengthened by more direct evidence, such as longitudinal body weight data, water intake, or comparison with group-housed no-task controls.

    (3) At the experimental level, the manuscript would also benefit from a more detailed characterization of training performance. Although three behavioral paradigms are presented, the data currently shown provide a stronger demonstration of feasibility than of training optimization. For a study focused on automated cognitive training, it would be critical to include more information on learning speed, progression across stages, success and failure rates, and variability across animals. Along the same lines, the comparison with manual training is a useful addition, but a broader benchmark including learning curves, time to criterion, and between-animal variability would make the practical value of the system easier to assess.

    (4) The authors claimed that they conducted 3 complex cognitive tasks (3AFC, 2AFC, 2AB) in their setup. However, those 3 tasks are quite basic for rodents and have been demonstrated in many studies, especially comparing tasks implemented in Yu et al., eLife 2025. Therefore, lowering this 'complex' statement is necessary.

    (5) The authors claimed that they have successfully implemented the so-called hybrid mode, but it is only briefly described and not supported by citations or data. Since this may be one of the most broadly applicable use cases of the platform, a more detailed explanation of how the system can be integrated with recording workflows would strengthen the manuscript.

    (6) The manuscript highlights the opportunity to relate task behavior to home-cage activity and to study individualized behavioral patterns. To better support these aspects, it would be helpful to include more subject-level analyses, rather than relying predominantly on population averages, or alternatively to discuss in more concrete terms which features of the dataset may be especially informative for studying individuality. More generally, the manuscript would benefit from clarifying whether different parameter settings within this group-housed framework may be better suited for maximizing training efficiency versus preserving more naturalistic or socially modulated behavior, and what the implications of these choices may be for interpretation.

    (7) In Table S1, 'Touch screen' is task-specific and is not necessarily a metric. 'Testing outside home cage' is also not necessarily an advantage (please clarify if it is). Many other systems implemented different levels of 'Alarm system', which is not reflected in the table.

    (8) Table S3 shows important data that help the reader to evaluate the paper's work, thus is deserved to move to the main text.

  3. Reviewer #2 (Public review):

    Summary:

    The Training Village (TV) is an innovative autonomous system for rodent training. By integrating an operant box with a group-housed home-cage environment, this platform enables animals to learn operant behaviors while preserving their social context and interactions, which is an aspect often overlooked in the field. The flexibility and modularity of the TV system allow training across multiple cognitive tasks in a continual learning framework. Furthermore, its remote accessibility and affordability make it a compelling tool for the broader neuroscience community.

    Comments:

    (1) Social Hierarchy and Access Competition

    Previous studies on rodent social hierarchy (e.g., PMID: 21960531) have demonstrated clear dominance structures within group-housed animals. Based on this, one might expect dominant animal(s) to occupy more sessions and trials than subordinate animals by preferentially accessing the operant box. Therefore, it is somewhat surprising to observe a relatively uniform distribution of operant box occupancy across animals (Figure 2a, 2i). As a control, it would strengthen the manuscript to include an independent assessment of social hierarchy (e.g., tube test, barber assay, or similar behavioral metrics) to quantitatively characterize dominance relationships within the cohort. Correlating these rankings with chamber occupancy and trial frequency would significantly strengthen the validation of the system's equity.

    (2) Behavioral Saving Effects in Continual Learning

    The authors demonstrate that the TV platform allows for the sequential learning of multiple cognitive tasks (Figure S3e). This provides an excellent opportunity to examine a continual learning paradigm. A key hallmark of successful continual learning is the "behavior savings effect", where re-learning a previously acquired task occurs faster than initial learning. For example, if animals are trained sequentially on task A (e.g., 2AFC), then task B (e.g., 2AB), and subsequently re-trained on task A, do they exhibit accelerated re-learning? Including such an analysis would significantly strengthen the claim regarding continual learning capabilities.

    (3) Robustness of Multi-Animal Attempt Detection

    In the TV platform, only one animal can access the operant box at a time under group-housed conditions. This setup inherently introduces the possibility of "multi-animal attempts", as shown in Figure 2j-k and Figure S2c. While the authors address this using pixel-based classification, additional quantitative validation would improve confidence in this approach. For instance, presenting the distribution of pixel counts for single-animal versus multi-animal events would be informative. Moreover, given variability in body size across animals, a fixed pixel threshold may not be sufficient. It would be helpful to include analyses of classification performance (e.g., Type I and Type II error rates) across different animal pairings within the same cohort.

    (4) Protocol Flexibility and Implementation

    It would be helpful to clarify how behavioral task protocols are switched within the TV system. Specifically, are task changes applied globally to all animals sharing the operant box, or can they be assigned individually? Additionally, are task sequences pre-programmed prior to the experiment, or can they be modified dynamically during ongoing experiments?

    (5) Presentation and Readability

    To improve readability, the Discussion section could be streamlined, as it is currently somewhat lengthy and descriptive.

  4. Reviewer #3 (Public review):

    Summary:

    The Training Village (TV) is an open-source automated platform for continuous training and testing of group-housed mice and rats in cognitive tasks. Animals live in enriched multi-compartment home cages and access a single operant box individually through a sorting corridor controlled by RFID identification and real-time video analysis. A Raspberry Pi 5 runs the entire system, manages an adaptive training algorithm, monitors animal welfare, and allows remote supervision via a graphical interface and Telegram alarm system. The system is validated across 12 groups totaling 121 animals, three cognitive paradigms of varying complexity, and experiments lasting up to 12 months.

    Strengths:

    (1) The open-source implementation is probably the paper's strongest point. The authors provide not just code but 3D-printable designs, a full bill of materials with costs (~5500€ total), assembly instructions, and a dedicated website. The estimated build time of 2-7 days is credible. In the current landscape of methods papers, this level of documentation is the minimum necessary to allow other laboratories to actually adopt and propagate the system - and the authors deliver it fully. The compatibility with two operant box designs, three cognitively distinct tasks, and two species - demonstrated empirically rather than merely claimed - makes the modularity argument credible and distinguishes the TV from systems designed around a single paradigm. Finally, the combination of automatic weighing at each exit, temperature and humidity tracking, and a granular Telegram alarm system (Table S2) represents a meaningful practical contribution. For a system operating 24/7 without daily human supervision, this level of welfare monitoring is a necessity, and it seems well implemented here.

    (2) With 121 animals across 12 groups, three distinct cognitive paradigms, two species, and longitudinal data spanning up to 12 months, the validation effort is substantial. The authors acknowledge the limitations of their comparisons - notably that the TV vs. manual training comparison is not a controlled experiment. The rat dataset is limited in scope, but the authors at least demonstrate that the system can be adapted to a second species, which is a useful proof of concept. The demonstration that task engagement increases progressively over 12 months (Fig. 3g) is a novel observation at this temporal scale, with practical implications for the design of long-term experiments.

    (3) The demonstration that operant box usage is distributed nearly uniformly across animals (Gini < 0.15 in all groups) is carefully demonstrated and addresses a question that any laboratory considering this type of system will legitimately ask, e.g., whether dominant individuals monopolize access at the expense of subordinates. This has been shown before in comparable systems, but remains a necessary validation for each new implementation. The control condition removing temporal constraints (Figure S4) adds useful mechanistic insight into the role of the refractory interval. However, the interpretation of this result deserves more nuance than the authors provide - see Weaknesses.

    Weaknesses:

    (1) The TV is more than an automation tool; its architecture makes the most sense if one intends to study how spontaneous home cage behavior relates to individual cognitive performance, and the introduction and discussion explicitly frame this as a key application. Yet the analysis delivers only group-level descriptive results, and the cognitive data are presented almost exclusively as group averages. The individual-level questions that the system is uniquely positioned to address (do stable home cage behavioral profiles emerge across animals, do animals learn at the same rate and using the same strategies, and do these dimensions correlate with each other ) are never asked. This is particularly relevant given that enriched social environments are precisely the conditions under which stable inter-individual differences tend to emerge spontaneously, even among genetically identical animals (Freund et al., 2013, Science), and that comparable systems have already linked such profiles to cognitive and neurochemical phenotypes (Torquet et al., 2018, Nature Communications). The TV clearly has the data to begin exploring this - doing so would substantially strengthen the paper's scientific contribution beyond its methodological value.

    (2) Sustained daytime operant box usage in nocturnal animals deserves more discussion: Box occupancy during the light phase remains around 75% - only modestly below the ~85% seen at night (Fig. S5a-b). The authors conclude this reflects "sustained engagement with the task throughout the circadian cycle," but other explanations are not considered: residual thirst driving animals to seek sucrose water during the day, and the refractory interval mechanically redistributing sessions into the light phase? A more explicit discussion of the consequences of 24/7 unsupervised testing for data quality (daytime sessions may yield noisier behavioral data?) would be useful.

    (3) The finding that all animals access the operant box in roughly equal proportions (Gini < 0.15) is practically important and carefully demonstrated. However, the authors' interpretation that animals self-organize in an egalitarian manner despite known social hierarchies deserves a note of caution. The system design itself constrains monopolization: the refractory interval imposes the same waiting time on all animals regardless of social rank, and session duration determines how often the box becomes available. The no-constraint control (Figure S4) partially addresses this but was run on already-trained animals, limiting its interpretive value. The key practical message, that all animals can access the task regularly under the proposed design, is well supported. Whether this reflects genuine social tolerance or is primarily a consequence of system constraints is a subtler question that the current data cannot fully resolve.

    (4) The rat cohort consists of a single group of 6 female Long-Evans rats, yet species comparisons are drawn across multiple dimensions (daily sessions, task engagement, performance...). Observed differences could reflect group size, sex, strain, reward calibration, or simple individual variability rather than species differences. These results should be presented for what they are: a useful proof of concept showing the system works with a second species, not a basis for comparative conclusions.