A Dual-Validation Framework for Temporal Robustness Assessment in Brain-Computer Interfaces for Motor Imagery

Read the full article See related articles

Listed in

This article is not in any list yet, why not save it to one of your lists.
Log in to save this article

Abstract

Brain–computer interfaces using motor imagery (MI-BCIs) offer a promising noninvasive communication pathway between humans and engineered equipment such as robots. However, for MI-BCIs based on electro encephalography (EEG) the reliability of the interface across recording sessions is limited by temporal non-stationary effects. Overcoming this barrier is critical to translating MI‑BCIs from controlled laboratory environments to practical uses. In this paper we present a novel dual-validation framework to rigorously evaluate temporal robustness of EEG signals of a MI-BCI. We collected data from six participants performing four motor imagery tasks using their hands and feet, namely, left/right hand clench and left/right foot plantar flexion. Features were extracted using Common Spatial Patterns, and ten machine learning classifiers were assessed within a unified pipeline. Our method integrates within-session evaluation (stratified K-fold cross-validation) with cross-session testing (bidirectional train/test), complemented by multi-dimensional stability metrics and performance heterogeneity assessment. Findings reveal minimal performance loss between conditions, with an average accuracy drop of just 2.5%. Our AdaBoost classifier achieved the highest within-session performance (84.0% system accuracy, F1-score: 83.8%/80.9% for hand/foot), while K-nearest neighbors (KNN) classifier demonstrated the greatest cross-session robustness (81.2% system accuracy, F1-score: 80.5%/80.2% for hand/foot, 0.663 robustness score). This study shows that robust performance across sessions is attainable for MI‑BCI evaluation, supporting the pathway toward reliable, real-world clinical deployment.

Article activity feed