Back

A graph-based deep learning framework for diabetic retinopathy classification with topology-aware feature augmentation

Belhadj, N. B.; Mezghich, M. A.; Fattahi, J.; Ghayoula, R.; Latrach, L.

2026-03-23 bioengineering
10.64898/2026.03.19.713075 bioRxiv
Show abstract

Diabetic retinopathy (DR) is the leading cause of preventable blindness in working-age adults, affecting an estimated 103 million people worldwide. Standard deep learning classifiers treat fundus images as independent samples, ignoring latent inter-patient relational structure that is most informative at clinically ambiguous intermediate severity levels. We propose a topology-aware, graph-based deep learning framework combining three complementary components: (i) an EfficientNet-B3 convolutional backbone for high-level visual feature extraction; (ii) persistent homology descriptors (H0 and H1) derived from morphologically skeletonised retinal vascular networks, characterising global vascular topology in a noise-robust manner; and (iii) a GraphSAGE graph neural network propagating disease-related information across a population-level similarity graph, refining representations through inductive neighbourhood aggregation. The similarity graph combines cosine similarity on visual features with 2-Wasserstein distance between persistence diagrams. Evaluated on three public benchmarks, the framework achieves 95.5% accuracy on Kaggle DR, 96.1% on Messidor-2, and 94.6% on APTOS 2019, consistently outperforming a strong CNN baseline by 1.5-2.3 percentage points across accuracy, Quadratic Weighted Kappa, and macro-F1. Ablation experiments confirm synergistic contributions of topological feature augmentation and relational graph learning. One-way ANOVA (F > 80, p < 0.001) confirms that DR progression is reflected in global vascular topology across all five severity stages, providing quantitative biological grounding for the framework design. Code and data are publicly available at https://github.com/Nader-BelHadj/plosene.

Matching journals

The top 5 journals account for 50% of the predicted probability mass.

1
Nature Communications
4913 papers in training set
Top 8%
17.3%
2
Nature Medicine
117 papers in training set
Top 0.1%
14.2%
3
Advanced Science
249 papers in training set
Top 2%
8.3%
4
npj Digital Medicine
97 papers in training set
Top 0.8%
6.2%
5
Nature Machine Intelligence
61 papers in training set
Top 0.5%
6.2%
50% of probability mass above
6
Communications Medicine
85 papers in training set
Top 0.1%
3.9%
7
Scientific Reports
3102 papers in training set
Top 38%
3.5%
8
Proceedings of the National Academy of Sciences
2130 papers in training set
Top 23%
3.0%
9
PLOS Computational Biology
1633 papers in training set
Top 11%
3.0%
10
Communications Biology
886 papers in training set
Top 5%
2.1%
11
Science Advances
1098 papers in training set
Top 15%
1.9%
12
EMBO Molecular Medicine
85 papers in training set
Top 1%
1.9%
13
Cell Systems
167 papers in training set
Top 7%
1.7%
14
PLOS ONE
4510 papers in training set
Top 55%
1.7%
15
PLOS Digital Health
91 papers in training set
Top 2%
1.2%
16
Nature Biomedical Engineering
42 papers in training set
Top 1%
1.2%
17
Cell Reports Medicine
140 papers in training set
Top 6%
1.1%
18
Medical Image Analysis
33 papers in training set
Top 0.8%
1.1%
19
Cell Reports
1338 papers in training set
Top 30%
0.9%
20
Science Translational Medicine
111 papers in training set
Top 5%
0.9%
21
Light: Science & Applications
16 papers in training set
Top 0.5%
0.9%
22
Cancer Research
116 papers in training set
Top 3%
0.9%
23
Bioinformatics
1061 papers in training set
Top 9%
0.9%
24
IEEE Transactions on Medical Imaging
18 papers in training set
Top 0.5%
0.7%
25
Journal of Vision
92 papers in training set
Top 0.5%
0.7%
26
Nature Methods
336 papers in training set
Top 6%
0.7%
27
iScience
1063 papers in training set
Top 38%
0.6%
28
Journal of The Royal Society Interface
189 papers in training set
Top 5%
0.6%