Back

More Signal vs. More Noise - Comparing Full Text and Abstract as Inputs for Large Language Model-based Classification of Oncology Trial Eligibility Criteria

Weyrich, J.; Dennstaedt, F.; Foerster, R.; Schroeder, C.; Aebersold, D. M.; Zwahlen, D. R.; Windisch, P.

2026-03-10 health informatics
10.64898/2026.03.10.26348010 medRxiv
Show abstract

PurposeLarge language models (LLMs) offer significant potential for automating the classification of clinical trials by eligibility criteria. However, a critical question remains regarding the optimal input data: while abstracts provide a condensed, high-density signal, full-text articles contain a much higher volume of information. It remains unclear whether the additional signal found in full texts improves classification performance or if the accompanying noise (in the form of thousands of words irrelevant to the question at hand in a complete manuscript) negatively affects the models reasoning capabilities. MethodsGPT-5 was applied to classify 200 randomized controlled oncology trials from high-impact medical journals, labeling them whether patients with localized and/or metastatic disease were eligible for inclusion. Each trial was classified twice - once using only the abstract and once using the full text - and GPT-5s outputs were compared with the ground-truth labels established by manual annotation. Performance was assessed by calculating and comparing accuracy, precision, recall, and F1 score, and the McNemar test was used to assess the statistical significance of the differences between the two input formats. ResultsFor identifying trials including patients with localized disease, GPT-5 achieved an accuracy of 86% (95% CI: 81% - 91%; F1 = 0.90) when using abstracts and 92% (95% CI: 88% - 95%; F1 = 0.92) when using full texts (p = 0.027). Performance for detecting trials, which include patients with metastatic disease, was comparably high, with accuracies of 99% (95% CI: 99% - 100%; F1 = 1.00) based on abstracts and 98% (95% CI: 97% - 100%; F1 = 0.99) based on full texts. Overall accuracy for assigning combined labels per trial increased from 86% (95% CI: 81% - 91%) using abstracts to 92% (95% CI: 88% - 95%) using full texts (p = 0.027). ConclusionProviding full-text articles to GPT-5 significantly improved the classification of trial eligibility criteria. These findings suggest that, for this task, the benefit of the additional signal contained within the full text outweighed the potential for performance degradation caused by increased noise. Utilizing full-text analysis appears particularly valuable for extracting specific eligibility criteria in oncology that are frequently omitted or not explicitly described within the abstract.

Matching journals

The top 8 journals account for 50% of the predicted probability mass.

1
JCO Clinical Cancer Informatics
18 papers in training set
Top 0.1%
17.2%
2
Cancer Medicine
24 papers in training set
Top 0.1%
7.0%
3
BMC Medical Informatics and Decision Making
39 papers in training set
Top 0.3%
7.0%
4
Artificial Intelligence in Medicine
15 papers in training set
Top 0.1%
6.2%
5
JMIR Medical Informatics
17 papers in training set
Top 0.2%
4.2%
6
Journal of Medical Internet Research
85 papers in training set
Top 1%
4.2%
7
BMC Medical Research Methodology
43 papers in training set
Top 0.2%
3.9%
8
PLOS ONE
4510 papers in training set
Top 41%
3.5%
50% of probability mass above
9
International Journal of Medical Informatics
25 papers in training set
Top 0.5%
3.2%
10
Scientific Reports
3102 papers in training set
Top 44%
2.7%
11
Journal of the American Medical Informatics Association
61 papers in training set
Top 0.9%
2.7%
12
Frontiers in Artificial Intelligence
18 papers in training set
Top 0.1%
2.7%
13
BMJ Health & Care Informatics
13 papers in training set
Top 0.2%
2.6%
14
JAMIA Open
37 papers in training set
Top 0.6%
2.3%
15
BMC Bioinformatics
383 papers in training set
Top 4%
1.9%
16
PLOS Digital Health
91 papers in training set
Top 2%
1.7%
17
npj Digital Medicine
97 papers in training set
Top 2%
1.7%
18
Computer Methods and Programs in Biomedicine
27 papers in training set
Top 0.5%
1.3%
19
Informatics in Medicine Unlocked
21 papers in training set
Top 0.7%
1.2%
20
BMJ Open
554 papers in training set
Top 11%
1.2%
21
Biology Methods and Protocols
53 papers in training set
Top 2%
1.1%
22
Journal of Clinical Epidemiology
28 papers in training set
Top 0.4%
1.1%
23
PLOS Computational Biology
1633 papers in training set
Top 21%
1.1%
24
Research Synthesis Methods
20 papers in training set
Top 0.2%
0.9%
25
BMC Infectious Diseases
118 papers in training set
Top 4%
0.9%
26
Database
51 papers in training set
Top 0.9%
0.8%
27
European Journal of Cancer
10 papers in training set
Top 0.5%
0.8%
28
Computers in Biology and Medicine
120 papers in training set
Top 4%
0.8%
29
Frontiers in Digital Health
20 papers in training set
Top 1%
0.8%
30
JCO Precision Oncology
14 papers in training set
Top 0.4%
0.8%