Multimodal EEG-fNIRS Fusion for Passive BCI-based Depressive State Classification
Sakurai, R.; Kojima, S.; Otake-Matsuura, M.; Kanoh, S.; Rutkowski, T. M.
Show abstract
Traditional psychiatric assessments for depression are often hindered by subjective bias and patient recall in-accuracy. This paper presents a multimodal passive Brain-Computer Interface (pBCI) designed for the objective screening of depressive traits through the end-to-end decoding of neural dynamics. We implemented a hybrid EEG-fNIRS framework to capture synchronized electro-hemodynamic responses during an emotional working memory (EWM) task. To classify sub-clinical depressive tendencies based on BDI-II scores, we utilized SincShallowNet, a deep learning architecture optimized for raw signal processing via learnable Sinc-filters. Our results demonstrate that the pBCI achieves peak performance in the auditory modality, with the integration of EEG and low-pass filtered fNIRS (0.15 Hz) yielding a balanced accuracy of 90.9% and an F1-score of 0.867. By isolating purely endogenous neural markers during the EWM maintenance phase, the system provides a robust "silent observer" for mental state monitoring. These findings validate the potential of multimodal pBCIs as high-precision, data-driven tools for early-stage depression screening, offering a scalable alternative to traditional clinical interviews and a foundation for longitudinal mental health monitoring.
Matching journals
The top 6 journals account for 50% of the predicted probability mass.