Back

Prediction models with survival data: a comparison between machine learning and the Cox proportional hazards model

Hazewinkel, A.-D.; Gelderblom, H.; Fiocco, M.

2022-04-02 oncology
10.1101/2022.03.29.22273112 medRxiv
Show abstract

Recent years have seen increased interest in using machine learning (ML) methods for survival prediction, chiefly using big datasets with mixed datatypes and/or many predictors Model comparisons have frequently been limited to performance measure evaluation, with the chosen measure often suboptimal for assessing survival predictive performance. We investigated ML model performance in an application to osteosarcoma data from the EURAMOS-1 clinical trial (NCT00134030). We compared the performance of survival neural networks (SNN), random survival forests (RSF) and the Cox proportional hazards model. Three performance measures suitable for assessing survival model predictive performance were considered: the C-index, and the time-dependent Brier and Kullback-Leibler scores. Comparisons were also made on predictor importance and patient-specific survival predictions. Additionally, the effect of ML model hyper-parameters on performance was investigated. All three models had comparable performance as assessed by the C-index and Brier and Kullback-Leibler scores, with the Cox model and SNN also comparable in terms of relative predictor importance and patient-specific survival predictions. RSFs showed a tendency for according less importance to predictors with uneven class distributions and predicting clustered survival curves, the latter a result of tuning hyperparameters that influence forest shape through restrictions on terminal node size and tree depth. SNNs were comparatively more sensitive to hyperparameter misspecification, with decreased regularization resulting in inconsistent predicted survival probabilities. We caution against using RSF for predicting patient-specific survival, as standard model tuning practices may result in aggregated predictions, which is not reflected in performance measure values, and recommend performing multiple reruns of SNNs to verify prediction consistency.

Matching journals

The top 7 journals account for 50% of the predicted probability mass.

1
Scientific Reports
3102 papers in training set
Top 4%
12.5%
2
PLOS ONE
4510 papers in training set
Top 16%
12.3%
3
Frontiers in Bioinformatics
45 papers in training set
Top 0.1%
6.4%
4
PLOS Computational Biology
1633 papers in training set
Top 6%
6.3%
5
Biology Methods and Protocols
53 papers in training set
Top 0.1%
6.3%
6
JCO Clinical Cancer Informatics
18 papers in training set
Top 0.1%
4.8%
7
JCO Precision Oncology
14 papers in training set
Top 0.1%
4.8%
50% of probability mass above
8
Cancers
200 papers in training set
Top 1%
4.8%
9
Frontiers in Artificial Intelligence
18 papers in training set
Top 0.1%
2.7%
10
Artificial Intelligence in Medicine
15 papers in training set
Top 0.2%
2.6%
11
Nature Communications
4913 papers in training set
Top 45%
2.4%
12
Frontiers in Oncology
95 papers in training set
Top 2%
2.1%
13
PeerJ
261 papers in training set
Top 6%
1.9%
14
eLife
5422 papers in training set
Top 42%
1.7%
15
European Journal of Cancer
10 papers in training set
Top 0.2%
1.7%
16
BMC Bioinformatics
383 papers in training set
Top 5%
1.7%
17
BMC Medical Research Methodology
43 papers in training set
Top 0.7%
1.5%
18
Breast Cancer Research
32 papers in training set
Top 0.4%
0.9%
19
Computer Methods and Programs in Biomedicine
27 papers in training set
Top 0.7%
0.9%
20
NAR Genomics and Bioinformatics
214 papers in training set
Top 3%
0.9%
21
Briefings in Bioinformatics
326 papers in training set
Top 6%
0.9%
22
Medical Physics
14 papers in training set
Top 0.5%
0.9%
23
Annals of Biomedical Engineering
34 papers in training set
Top 1%
0.8%
24
iScience
1063 papers in training set
Top 29%
0.8%
25
Cancer Research Communications
46 papers in training set
Top 1%
0.7%
26
Statistics in Medicine
34 papers in training set
Top 0.4%
0.7%
27
Frontiers in Genetics
197 papers in training set
Top 11%
0.7%
28
GigaScience
172 papers in training set
Top 3%
0.7%
29
IEEE/ACM Transactions on Computational Biology and Bioinformatics
32 papers in training set
Top 0.7%
0.7%
30
Journal of Medical Internet Research
85 papers in training set
Top 5%
0.7%