Back

C-RLM: Schema-Enforced Recursive Synthesis for Auditable, Long-Context Clinical Documentation

Yu, Y.

2026-01-26 health informatics
10.64898/2026.01.24.26344761 medRxiv
Show abstract

Clinical decision-making for multi-morbid patients requires synthesizing evidence from lengthy, fragmented records--a task that exposes the limitations of standard Retrieval-Augmented Generation (RAG) and long-context Large Language Models (LLMs), which often lose critical information or lack auditability. We introduce the Clinical-Recursive Language Model (C-RLM), a framework that reframes evidence synthesis as a structured, recursive compilation process rather than a single-pass retrieval task. C-RLM iteratively builds a validated knowledge state using schema-enforced transitions, a Robust Nomenclature Resilience (RNR) layer for synonym consolidation, and a TraceTracker system for deterministic provenance. Evaluated on 100 complex Lupus Nephritis case reports ([~]24.5k tokens each), C-RLM achieves 100% structural consistency and 99% regimen recall (F1), outperforming a strong Flat RAG baseline. While introducing a 2.7x computational overhead, C-RLM delivers a crucial "Synthesis Dividend": recovery of clinically critical entities fragmented across distant text spans, with full auditability back to source text offsets. Our results demonstrate that for safety-critical clinical applications, the trade-off in latency is justified by gains in reliability, auditability, and support for human-in-the-loop governance.

Matching journals

The top 6 journals account for 50% of the predicted probability mass.

1
npj Digital Medicine
97 papers in training set
Top 0.2%
23.4%
2
Journal of the American Medical Informatics Association
61 papers in training set
Top 0.3%
8.7%
3
Bioinformatics
1061 papers in training set
Top 3%
7.1%
4
Nature Communications
4913 papers in training set
Top 27%
6.6%
5
Med
38 papers in training set
Top 0.1%
4.1%
6
JCO Clinical Cancer Informatics
18 papers in training set
Top 0.2%
3.7%
50% of probability mass above
7
Scientific Reports
3102 papers in training set
Top 33%
3.7%
8
iScience
1063 papers in training set
Top 4%
3.7%
9
PLOS ONE
4510 papers in training set
Top 43%
2.8%
10
Nature Medicine
117 papers in training set
Top 1%
2.7%
11
Science Translational Medicine
111 papers in training set
Top 2%
2.0%
12
European Respiratory Journal
54 papers in training set
Top 0.9%
1.8%
13
Nature Computational Science
50 papers in training set
Top 0.7%
1.5%
14
The Lancet Digital Health
25 papers in training set
Top 0.5%
1.5%
15
Patterns
70 papers in training set
Top 1%
1.3%
16
Nature Machine Intelligence
61 papers in training set
Top 3%
1.0%
17
PLOS Digital Health
91 papers in training set
Top 2%
1.0%
18
eLife
5422 papers in training set
Top 52%
0.9%
19
Advanced Science
249 papers in training set
Top 16%
0.9%
20
Cell Systems
167 papers in training set
Top 11%
0.8%
21
Nature Biotechnology
147 papers in training set
Top 7%
0.8%
22
Journal of Biomedical Informatics
45 papers in training set
Top 1%
0.8%
23
Nature Biomedical Engineering
42 papers in training set
Top 2%
0.8%
24
Philosophical Transactions of the Royal Society B
51 papers in training set
Top 5%
0.8%
25
Frontiers in Digital Health
20 papers in training set
Top 2%
0.7%
26
Communications Medicine
85 papers in training set
Top 2%
0.5%