Back

Leveraging large language models to address common vaccination myths and misconceptions

Reis, F.; Bayer, L. J.; Malerczyk, C.; Lenz, C.; von Eiff, C.

2026-03-02 health informatics
10.64898/2026.02.27.26347254 medRxiv
Show abstract

Large language models (LLMs) are increasingly used by the public to seek health information, yet their reliability in addressing common vaccine myths remains unclear. We conducted an exploratory multi-vendor evaluation of three LLMs (GPT-5, Gemini 2.5 Flash, Claude Sonnet 4) using officially curated vaccination myths from Germanys public health institution and two realistic user framings as prompts: a curious skeptic and a convinced believer. All model responses were independently evaluated by two blinded medical experts for misconception addressal (binary), scientific accuracy, and communication clarity (5-point Likert scales). Additionally, blinded marketing experts ranked models for lay communication clarity, and Flesch-Kincaid Reading Ease scores were computed for all outputs. Across all myths, prompts, and models (11 x 2 x 3 = 66 rating items), medical raters found 100% successful refutation of misinformation. Scientific accuracy and clarity ratings were high and tightly clustered (median 4.0-4.5), with no combined score below 3 and substantial inter-rater agreement. Marketing experts independently ranked Gemini 2.5 Flash and GPT-5 highest for lay clarity, with Claude Sonnet 4 consistently less favored. Readability analysis revealed generally low accessibility, particularly for the convinced believer framing and for Claude Sonnet 4 outputs. Our findings suggest that current general-purpose LLMs can deliver accurate debunking of widely documented vaccine myths under realistic conditions, but that linguistic complexity and framing-sensitive style may limit accessibility. Careful integration of LLMs into public health channels, alongside transparent sourcing and readability optimization, could enable these models to be used as scalable tools for debunking vaccine myths.

Matching journals

The top 8 journals account for 50% of the predicted probability mass.

1
npj Digital Medicine
97 papers in training set
Top 0.4%
12.4%
2
Journal of Medical Internet Research
85 papers in training set
Top 0.4%
10.1%
3
JMIR Public Health and Surveillance
45 papers in training set
Top 0.1%
6.8%
4
International Journal of Medical Informatics
25 papers in training set
Top 0.2%
6.4%
5
PLOS Digital Health
91 papers in training set
Top 0.5%
4.9%
6
Scientific Reports
3102 papers in training set
Top 24%
4.9%
7
Scientific Data
174 papers in training set
Top 0.4%
4.0%
8
Computers in Biology and Medicine
120 papers in training set
Top 0.8%
3.7%
50% of probability mass above
9
Journal of the American Medical Informatics Association
61 papers in training set
Top 0.9%
3.1%
10
Frontiers in Digital Health
20 papers in training set
Top 0.4%
2.5%
11
PLOS ONE
4510 papers in training set
Top 46%
2.4%
12
Healthcare
16 papers in training set
Top 0.3%
2.4%
13
JAMIA Open
37 papers in training set
Top 0.7%
1.9%
14
BMC Medical Informatics and Decision Making
39 papers in training set
Top 1%
1.9%
15
BMJ Health & Care Informatics
13 papers in training set
Top 0.3%
1.9%
16
Frontiers in Public Health
140 papers in training set
Top 5%
1.7%
17
GigaScience
172 papers in training set
Top 1%
1.7%
18
Philosophical Transactions of the Royal Society B
51 papers in training set
Top 3%
1.5%
19
BMC Infectious Diseases
118 papers in training set
Top 3%
1.3%
20
Computational and Structural Biotechnology Journal
216 papers in training set
Top 6%
1.2%
21
Journal of Biomedical Informatics
45 papers in training set
Top 1%
1.1%
22
Nature Human Behaviour
85 papers in training set
Top 3%
1.0%
23
Vaccine
189 papers in training set
Top 2%
0.9%
24
Biology Methods and Protocols
53 papers in training set
Top 2%
0.9%
25
BMC Medical Research Methodology
43 papers in training set
Top 1%
0.8%
26
Computer Methods and Programs in Biomedicine
27 papers in training set
Top 1.0%
0.7%
27
BMC Bioinformatics
383 papers in training set
Top 7%
0.7%
28
JMIR Medical Informatics
17 papers in training set
Top 1%
0.7%
29
The Lancet Digital Health
25 papers in training set
Top 1%
0.7%
30
DIGITAL HEALTH
12 papers in training set
Top 0.7%
0.7%