Back

An Empirical Assessment of Inferential Reproducibility of Linear Regression in Health and Biomedical Research Papers

Jones, L.; Barnett, A.; Hartel, G.; Vagenas, D.

2026-04-07 health systems and quality improvement
10.64898/2026.04.07.26350296 medRxiv
Show abstract

Background: In health research, variability in modelling decisions can lead to different conclusions even when the same data are analysed, a challenge known as inferential reproducibility. In linear regression analyses, incorrect handling of key assumptions, such as normality of the residuals and linearity, can undermine reproducibility. This study examines how violations of these assumptions influence inferential conclusions when the same data are reanalysed. Methods: We randomly sampled 95 health-related PLOS ONE papers from 2019 that reported linear regression in their methods. Data were available for 43 papers, and 20 were assessed for computational reproducibility, with three models per paper evaluated. The 14 papers that included a model at least partially computationally reproduced were then examined for inferential reproducibility. To assess the impact of assumption violations, differences in coefficients, 95% confidence intervals, and model fit were compared. Results: Of the fourteen papers assessed, only three were inferentially reproducible. The most frequently violated assumptions were normality and independence, each occurring in eight papers. Violations of independence were particularly consequential and were commonly associated with inferential failure. Although reproduced analyses often retained the same binary statistical significance classification as the original studies, confidence intervals were frequently wider, indicating greater uncertainty and reduced precision. Such uncertainty may affect the interpretation of results and, in turn, influence treatment decisions and clinical practice. Conclusion: Our findings demonstrate that substantial violations of key modelling assumptions often went undetected by authors and peer reviewers and, in many cases, were associated with inferential reproducibility failure. This highlights the need for stronger statistical education and greater transparency in modelling decisions. Rather than applying rigid or misinformed rules, such as incorrectly testing the normality of the outcome variable, researchers should adopt modelling frameworks guided by the research question and the study design. When assumptions are violated, appropriate alternatives, such as robust methods, bootstrapping, generalized linear models, or mixed-effects models, should be considered. Given that assumption violations were common even in relatively simple regression models, early and sustained collaboration with statisticians is critical for supporting robust, defensible, and clinically meaningful conclusions.

Matching journals

The top 3 journals account for 50% of the predicted probability mass.

1
Journal of Clinical Epidemiology
28 papers in training set
Top 0.1%
33.4%
2
BMJ Open
554 papers in training set
Top 2%
10.2%
3
PLOS ONE
4510 papers in training set
Top 18%
10.2%
50% of probability mass above
4
Research Synthesis Methods
20 papers in training set
Top 0.1%
4.9%
5
Trials
25 papers in training set
Top 0.4%
3.6%
6
BMC Medical Research Methodology
43 papers in training set
Top 0.4%
2.1%
7
Journal of Biomedical Informatics
45 papers in training set
Top 0.7%
1.9%
8
PLOS Biology
408 papers in training set
Top 8%
1.8%
9
European Journal of Epidemiology
40 papers in training set
Top 0.3%
1.7%
10
Royal Society Open Science
193 papers in training set
Top 2%
1.7%
11
BMC Medicine
163 papers in training set
Top 4%
1.7%
12
Statistics in Medicine
34 papers in training set
Top 0.2%
1.7%
13
BMJ Global Health
98 papers in training set
Top 2%
1.5%
14
F1000Research
79 papers in training set
Top 3%
1.2%
15
Biometrics
22 papers in training set
Top 0.1%
1.2%
16
Computer Methods and Programs in Biomedicine
27 papers in training set
Top 0.6%
1.1%
17
JAMA Network Open
127 papers in training set
Top 3%
1.1%
18
Journal of Public Health
23 papers in training set
Top 0.8%
1.0%
19
International Journal of Epidemiology
74 papers in training set
Top 2%
1.0%
20
Medical Decision Making
10 papers in training set
Top 0.2%
0.9%
21
BMC Medical Informatics and Decision Making
39 papers in training set
Top 2%
0.8%
22
Medicine
30 papers in training set
Top 2%
0.8%
23
Journal of the American Medical Informatics Association
61 papers in training set
Top 2%
0.8%
24
BMJ Open Quality
15 papers in training set
Top 0.9%
0.7%
25
Canadian Medical Association Journal
15 papers in training set
Top 0.4%
0.7%
26
PLOS Digital Health
91 papers in training set
Top 3%
0.7%
27
JMIRx Med
31 papers in training set
Top 3%
0.5%
28
Heart
10 papers in training set
Top 1%
0.5%
29
Scientific Reports
3102 papers in training set
Top 80%
0.5%