A Dual-Validation Framework for Temporal Robustness Assessment in Brain-Computer Interfaces for Motor Imagery
Listed in
This article is not in any list yet, why not save it to one of your lists.Abstract
Brain–computer interfaces using motor imagery (MI-BCIs) offer a promising noninvasive communication pathway between humans and engineered equipment such as robots. However, for MI-BCIs based on electro encephalography (EEG) the reliability of the interface across recording sessions is limited by temporal non-stationary effects. Overcoming this barrier is critical to translating MI‑BCIs from controlled laboratory environments to practical uses. In this paper we present a novel dual-validation framework to rigorously evaluate temporal robustness of EEG signals of a MI-BCI. We collected data from six participants performing four motor imagery tasks using their hands and feet, namely, left/right hand clench and left/right foot plantar flexion. Features were extracted using Common Spatial Patterns, and ten machine learning classifiers were assessed within a unified pipeline. Our method integrates within-session evaluation (stratified K-fold cross-validation) with cross-session testing (bidirectional train/test), complemented by multi-dimensional stability metrics and performance heterogeneity assessment. Findings reveal minimal performance loss between conditions, with an average accuracy drop of just 2.5%. Our AdaBoost classifier achieved the highest within-session performance (84.0% system accuracy, F1-score: 83.8%/80.9% for hand/foot), while K-nearest neighbors (KNN) classifier demonstrated the greatest cross-session robustness (81.2% system accuracy, F1-score: 80.5%/80.2% for hand/foot, 0.663 robustness score). This study shows that robust performance across sessions is attainable for MI‑BCI evaluation, supporting the pathway toward reliable, real-world clinical deployment.