From video to behaviour: an LSTM-based approach for automated nest behaviour recognition in the wild
Discuss this preprint
Start a discussion What are Sciety discussions?Listed in
This article is not in any list yet, why not save it to one of your lists.Abstract
Studies of animal behaviour usually rely on direct observations or manual annotations of video recordings. However, such methods can be very time-consuming and error-prone, leading to sub-optimal sample sizes. Recent advances in deep-learning show great potential to overcome such limitations, however, most currently available behavioural recognition solutions remain focused on standardised settings.
Here, we present a deployment-focused framework to guide researchers in building behavioural recognition systems from video data, using Long Short-Term Memory (LSTM) networks to classify behavioural sequences across consecutive frames.
LSTMs allowed us to: 1) monitor nest activity by detecting the birds’ presence and simultaneously classifying the type of trajectory: i.e., nest-chamber entrance or exit; and 2) identify the behaviour performed: building, aggression or sanitation. Using our framework, we largely outperformed human annotators in error and speed. LSTM performance improved with challenging training instances, and remained robust even with modest sample sizes. LSTM also outperformed YOLO, highlighting the critical role of temporal sequence information in behavioural analysis.
We demonstrate that our approach is replicable across three bird species and applicable to deployment videos, highlighting its value as a generalizable and transferable tool for long-term studies in the wild.