C-RLM: Schema-Enforced Recursive Synthesis for Auditable, Long-Context Clinical Documentation
Yu, Y.
Show abstract
Clinical decision-making for multi-morbid patients requires synthesizing evidence from lengthy, fragmented records--a task that exposes the limitations of standard Retrieval-Augmented Generation (RAG) and long-context Large Language Models (LLMs), which often lose critical information or lack auditability. We introduce the Clinical-Recursive Language Model (C-RLM), a framework that reframes evidence synthesis as a structured, recursive compilation process rather than a single-pass retrieval task. C-RLM iteratively builds a validated knowledge state using schema-enforced transitions, a Robust Nomenclature Resilience (RNR) layer for synonym consolidation, and a TraceTracker system for deterministic provenance. Evaluated on 100 complex Lupus Nephritis case reports ([~]24.5k tokens each), C-RLM achieves 100% structural consistency and 99% regimen recall (F1), outperforming a strong Flat RAG baseline. While introducing a 2.7x computational overhead, C-RLM delivers a crucial "Synthesis Dividend": recovery of clinically critical entities fragmented across distant text spans, with full auditability back to source text offsets. Our results demonstrate that for safety-critical clinical applications, the trade-off in latency is justified by gains in reliability, auditability, and support for human-in-the-loop governance.
Matching journals
The top 6 journals account for 50% of the predicted probability mass.