Back

Evaluation of polygenic scoring methods in five biobanks reveals greater variability between biobanks than between methods and highlights benefits of ensemble learning

Monti, R.; Eick, L.; Hudjashov, G.; Läll, K.; Kanoni, S.; Wolford, B. N.; Wingfield, B.; Pain, O.; Wharrie, S.; Jermy, B.; McMahon, A.; Hartonen, T.; Heyne, H. O.; Mars, N.; Genes & Health Research Team, ; Hveem, K.; Inouye, M.; van Heel, D. A.; Mägi, R.; Marttinen, P.; Ripatti, S.; Ganna, A.; Lippert, C.

2023-11-20 genetic and genomic medicine
10.1101/2023.11.20.23298215 medRxiv
Show abstract

Methods to estimate polygenic scores (PGS) from genome-wide association studies are increasingly utilized. However, independent method evaluation is lacking, and method comparisons are often limited. Here, we evaluate polygenic scores derived using seven methods in five biobank studies (totaling about 1.2 million participants) across 16 diseases and quantitative traits, building on a reference-standardized framework. We conducted meta-analyses to quantify the effects of method choice, hyperparameter tuning, method ensembling and target biobank on PGS performance. We found that no single method consistently outperformed all others. PGS effect sizes were more variable between biobanks than between methods within biobanks when methods were well-tuned. Differences between methods were largest for the two investigated autoimmune diseases, seropositive rheumatoid arthritis and type 1 diabetes. For most methods, cross-validation was more reliable for tuning hyperparameters than automatic tuning (without the use of target data). For a given target phenotype, elastic net models combining PGS across methods (ensemble PGS) tuned in the UK Biobank provided consistent, high, and cross-biobank transferable performance, increasing PGS effect sizes ({beta}-coefficients) by a median of 5.0% relative to LDpred2 and MegaPRS (the two best performing single methods when tuned with cross-validation). Our interactively browsable online-results (https://methodscomparison.intervenegeneticscores.org/) and open-source workflow prspipe (https://github.com/intervene-EU-H2020/prspipe) provide a rich resource and reference for the analysis of polygenic scoring methods across biobanks.

Matching journals

The top 4 journals account for 50% of the predicted probability mass.

1
Cell Genomics
162 papers in training set
Top 0.1%
26.6%
2
Nature Communications
4913 papers in training set
Top 13%
13.0%
3
The American Journal of Human Genetics
206 papers in training set
Top 0.4%
10.4%
4
Briefings in Bioinformatics
326 papers in training set
Top 0.7%
7.0%
50% of probability mass above
5
Genome Medicine
154 papers in training set
Top 0.9%
6.5%
6
Frontiers in Genetics
197 papers in training set
Top 2%
3.7%
7
Human Genetics and Genomics Advances
70 papers in training set
Top 0.1%
3.7%
8
Nature Genetics
240 papers in training set
Top 3%
3.0%
9
Bioinformatics
1061 papers in training set
Top 6%
2.7%
10
Scientific Reports
3102 papers in training set
Top 56%
1.7%
11
Nucleic Acids Research
1128 papers in training set
Top 12%
1.5%
12
Communications Biology
886 papers in training set
Top 10%
1.5%
13
BMC Genomics
328 papers in training set
Top 4%
1.0%
14
NAR Genomics and Bioinformatics
214 papers in training set
Top 3%
0.8%
15
PLOS Computational Biology
1633 papers in training set
Top 23%
0.8%
16
Human Genomics
21 papers in training set
Top 0.3%
0.8%
17
Genome Biology
555 papers in training set
Top 7%
0.8%
18
BMC Medical Genomics
36 papers in training set
Top 1%
0.8%
19
eBioMedicine
130 papers in training set
Top 4%
0.7%
20
Human Genetics
25 papers in training set
Top 0.4%
0.7%
21
International Journal of Epidemiology
74 papers in training set
Top 3%
0.5%
22
eLife
5422 papers in training set
Top 63%
0.5%
23
Bioinformatics Advances
184 papers in training set
Top 5%
0.5%
24
The Lancet Rheumatology
11 papers in training set
Top 0.3%
0.5%