Back

Predicting Post-Stroke Aphasia Speech Performance from Multimodal Data with Explainable Machine Learning

Parchure, S.; Gupta, A.; Kelkar, A.; Vnenchak, L.; Faseyitan, O.; Medaglia, J. D.; Harvey, D. Y.; Coslett, H. B.; Hamilton, R. H.

2026-02-05 bioengineering
10.64898/2026.02.02.703416 bioRxiv
Show abstract

Aphasia, an acquired language deficit, is the most common post-stroke focal cognitive impairment, and roughly 60% cases become chronic (duration >6 months). Aphasia therapies could be optimized if clinicians could make personalized predictions of how individual persons with aphasia (PWA) would be likely to perform on particular language tasks. However, current approaches relying on imaging, lesion volume, patient demographics, and clinical scores achieve less than 50% accuracy in predicting performance in PWA. Research algorithms using complex imaging and fMRI can make binary predictions about the presence or absence of aphasia but do not give more clinically relevant information. We aim to predict word-by-word speech accuracy in PWA to better enable personalized speech therapies. To be clinically informative, machine learning models developed for this purpose should use clinically available inputs, explain key features behind a prediction, and generalize to new PWA and previously unseen words. This study combines multimodal input features from clinical testing scores and structural MRI neuroimaging with a novel data source: word-by-word linguistic difficulty. We computed metrics of cognitive burden, such as semantic selection and recall demands, and articulatory burden, such as word length in phonemes and syllables, using naturalistic corpora containing over a billion words of English text. Retrospective training, ten-fold cross validation and 500-run bootstrapping of different machine learning models with various combinations of input features was conducted using 4620 trials. A simplified version of the best model using widely available inputs was deployed clinically through a web app, and prospective generalization was tested on 570 trials with unseen words and different naming tasks in new PWA. We found the best performances with random forest classifiers using linguistic difficulty combined with either clinical information (AUROC {+/-} SEM = 0.87 {+/-} 0.07), or all together with structural imaging connectivity (0.90 {+/-} 0.04). Classifiers using multimodal inputs significantly outperformed others employing single inputs (range 0.66-0.85, p<0.05). Extracting feature importances from the best model showed that Western Aphasia Battery scores, semantic demands, number of phonemes, and syllables were predictive of PWA speech accuracy. Structural integrity in peri-lesional brain regions predicted better language performance whereas higher connectivity of select contralateral homotopes contributed to prediction of worse speech. Without the inclusion of MRI data, lesion volume was a key predictor of PWA speech as well. A simplified, clinically ready, explainable model (publicly available as AphasiaLENS web application) predicted PWA accuracy for any user-entered word, not restricted to a standardized battery. Its prospective generalization performance was not significantly different from the best model using full inputs (AUROC ranges 0.81-0.89, p>0.05). Thus, our research can help inform individualized treatment planning for PWA, while also suggesting research targets through better understanding of brain-behavior relationships.

Matching journals

The top 8 journals account for 50% of the predicted probability mass.

1
Scientific Reports
3102 papers in training set
Top 6%
10.2%
2
Brain
154 papers in training set
Top 0.7%
8.3%
3
Human Brain Mapping
295 papers in training set
Top 0.8%
7.2%
4
NeuroImage: Clinical
132 papers in training set
Top 0.6%
6.9%
5
PLOS ONE
4510 papers in training set
Top 25%
6.9%
6
Journal of NeuroEngineering and Rehabilitation
28 papers in training set
Top 0.2%
4.9%
7
Alzheimer's Research & Therapy
52 papers in training set
Top 0.5%
4.3%
8
Annals of Clinical and Translational Neurology
29 papers in training set
Top 0.3%
3.6%
50% of probability mass above
9
Frontiers in Neuroscience
223 papers in training set
Top 2%
2.8%
10
Journal of Neural Engineering
197 papers in training set
Top 0.9%
2.6%
11
European Radiology
14 papers in training set
Top 0.3%
2.5%
12
Computers in Biology and Medicine
120 papers in training set
Top 1%
2.4%
13
Brain Connectivity
22 papers in training set
Top 0.1%
2.1%
14
Brain Communications
147 papers in training set
Top 1%
2.1%
15
NeuroImage
813 papers in training set
Top 4%
1.8%
16
Frontiers in Neurology
91 papers in training set
Top 3%
1.7%
17
Communications Biology
886 papers in training set
Top 8%
1.7%
18
Neurobiology of Language
28 papers in training set
Top 0.1%
1.7%
19
Philosophical Transactions of the Royal Society B
51 papers in training set
Top 4%
1.2%
20
European Journal of Neuroscience
168 papers in training set
Top 0.8%
1.2%
21
Frontiers in Aging Neuroscience
67 papers in training set
Top 3%
1.0%
22
Proceedings of the National Academy of Sciences
2130 papers in training set
Top 40%
1.0%
23
Bioengineering
24 papers in training set
Top 1%
0.9%
24
PLOS Computational Biology
1633 papers in training set
Top 22%
0.9%
25
Journal of Speech, Language, and Hearing Research
10 papers in training set
Top 0.1%
0.8%
26
Neurobiology of Disease
134 papers in training set
Top 4%
0.8%
27
F1000Research
79 papers in training set
Top 5%
0.8%
28
Journal of Alzheimer’s Disease
39 papers in training set
Top 1%
0.7%
29
npj Digital Medicine
97 papers in training set
Top 4%
0.6%
30
Frontiers in Digital Health
20 papers in training set
Top 2%
0.6%