Back

ImmunoInformatics

Elsevier BV

Preprints posted in the last 7 days, ranked by how well they match ImmunoInformatics's content profile, based on 11 papers previously published here. The average preprint has a 0.01% match score for this journal, so anything above that is already an above-average fit.

1
Deciphering antigen-driven T cell responses through vectorized TCRdist sequence neighborhood quantification

Valkiers, S.; Mayer-Blackwell, K.; Yeh, A. C.; Van Deuren, V. M. L.; Fiore-Gartland, A.; Hill, G.; Laukens, K.; Meysman, P.; Bradley, P.

2026-04-14 immunology 10.64898/2026.04.10.717405 medRxiv
Top 2%
0.1%
Show abstract

T cells provide precise mechanisms to defend the body against infection and malignancies, mediated through the expression of their hypervariable T cell receptors (TCRs). Interpreting similarity between TCRs, however, remains a significant challenge. While performant clustering methods exist, these often fail to distinguish between antigen-driven convergent selection and patterns arising stochastically from biases in the V(D)J recombination mechanism. Moreover, defining enrichment in sequence similarity among large repertoires is computationally taxing. To address these limitations, we present an efficient computational framework for rapid approximation of TCRdist distances using fixed-length vector embeddings and highly optimized nearest neighbor search, allowing sequence similarity enrichment testing at a multi-repertoire-wide scale. This framework leverages a novel shuffling-based background model that preserves important repertoire characteristics such as V gene frequency, CDR3 sequence length and generation probability more accurately than synthetic models. Together, these tools enable the efficient and robust identification of significantly neighbor enriched (SNE) TCR sequences at scale. We validate this approach by showing a significant enrichment of SNE clones in memory T cell fractions and further demonstrate its utility in identifying convergent T cell signatures of response to vaccination and viral infections, providing a scalable approach for antigen-agnostic T cell response profiling.

2
Spatiotemporal transcriptomic analysis during cold ischemic injury to the murine kidney reveals compartment-specific changes

Singh, S.; Patel, S. K.; Matsuura, R.; Velazquez, D.; Sun, Z.; Noel, S.; Rabb, H.; Fan, J.

2026-04-18 bioinformatics 10.1101/2025.05.25.654911 medRxiv
Top 2%
0.0%
Show abstract

Background: Kidney transplantation is the preferred treatment strategy for end-stage kidney disease. Deceased donor kidneys usually undergo cold storage until kidney transplantation, leading to cold ischemia injury that may contribute to poor graft outcomes. However, the molecular characterization of potential mechanisms of cold ischemia injury remains incomplete. Results: To bridge this knowledge gap, we leveraged the 10x Visium spatial transcriptomic technology to perform full transcriptome profiling of murine kidneys subject to varying durations of cold ischemia typical in a deceased donor kidney transplant setting. We developed a computational workflow to identify and compare spatiotemporal transcriptomic changes that accompany the injury pathophysiology in a tissue compartment-specific manner. We identified proportional enrichment of oxidative phosphorylation (OXPHOS) genes with increasing duration of cold ischemia injury within the oxygen-lean inner medulla region, suggestive of atypical metabolic presentation. This was distinct in cold ischemia injury tissue compared to warm ischemia-reperfusion kidney injury tissue. Spatiotemporal trends were validated by qPCR and immunofluorescence in a larger cohort of mice. We provide an interactive online browser at https://jef.works/CellCarto-ColdIschemia/ to facilitate exploration of our results by the broader scientific and clinical community. Conclusions: Altogether, our spatiotemporal transcriptomic analysis identified coordinated molecular changes within metabolic pathways such as OXPHOS deep within the cold ischemic kidney, highlighting the need for increased attention to the inner medulla and potential opportunities for new insights beyond those available from superficial biopsy-focused tissue examinations.

3
DyME: An MD-based engine exploiting HTP mutagenesis for protein engineering and recognition mimicry

Guillem-Gloria, P. M.; Ruiz-Gomez, G.; Pisabarro, M. T.

2026-04-13 bioinformatics 10.64898/2026.04.10.717642 medRxiv
Top 2%
0.0%
Show abstract

Protein recognition mimicry is of great interest in the field of molecular bioengineering and rational design, with mutagenesis frequently employed to analyze the effects of altering amino acids involved in molecular recognition. The conformational and energetic effects of such alterations can be investigated in detail with the help of molecular dynamics (MD) methodologies. While existing MD-based computational tools can be used to explore a particular set of mutations at a time, suitable for small-scale studies, high-throughput (HTP) exploration of protein recognition for engineering purposes would greatly benefit from an integrative platform that streamlines preparation, mutagenesis, simulation and post-processing of up to several thousand molecular systems, along with robust tools for comprehensive and straightforward comparative analysis. DyME (Dynamic Mutagenesis Engine) is a distributed platform that enables systematic investigations of protein recognition mimicry by combining HTP mutagenesis, solvated MD simulations and a Toolbox for comparative analysis (TCA), including interfacial water-site mapping. DyME uses 3D structural information of any protein-protein or protein-DNA complex as input. Its automated MD-based mutagenesis engine facilitates systematic investigation of how site-specific alterations affect recognition, enabling the organization of single, double and triple modifications into combinatorial libraries for comprehensive comparative analysis. In DyME, relevant MD trajectory-derived data is scavenged and stored into a central database, providing aggregation capabilities that ease multi-feature analysis across an extensive collection of simulations. An interactive web-GUI and specialized widgets simplify preparation and efficient molecular and numerical comparative exploration. DyMEs capabilities are evaluated using available experimental data. Its source code is available at https://github.com/pisabarro-group/DYME

4
TB-Bench: A Systematic Benchmark of Machine Learning and Deep Learning Methods for Second-Line TB Drug Resistance Prediction

VP, B.; Jaiswal, S.; Meshram, A.; PVS, D.; S C, S.; Narayanan, M.

2026-04-13 bioinformatics 10.64898/2026.04.08.717138 medRxiv
Top 2%
0.0%
Show abstract

Drug-resistant tuberculosis (TB), characterized by prolonged treatment regimens and suboptimal treatment outcomes, remains a major obstacle to global TB elimination. Advances in sequencing technologies have enabled the development of machine-learning (ML) approaches, including deep-learning (DL) methods, to predict drug resistance directly from genomic data. However, a significant gap remains in translating these advances into clinical practice. While current approaches reliably predict resistance to first-line drugs, they show consistently lower and more variable performance for second-line drugs compared with traditional drug-susceptibility testing. To characterize these limitations and assess practical utility, we conducted a comprehensive survey and standardized benchmarking of current approaches for predicting TB drug resistance using whole-genome sequencing (WGS) data. Using systematic selection criteria, we identified 20 traditional ML and DL models from 8 studies and evaluated drug-specific versions across 14 second-line drugs within a unified framework. To account for methodological heterogeneity, the models were evaluated using three distinct feature sets reflecting variability in input representations. We trained and evaluated the models on different subsets of the WHO dataset, comprising 50,801 samples, and assessed generalizability using an external validation dataset comprising 1,199 samples. In the internal evaluation on the held-out WHO test dataset, traditional ML models using binary features achieved higher predictive performance than DL models. For example, XGBoost achieved the highest area under the precision-recall curve (PRAUC) scores (46%-93%) for 10 of the 14 drugs. However, performance varied substantially across drugs. Notably, the superior performance of traditional ML models -- even with limited feature sets -- highlights their applicability in low-resource settings. When evaluated on the external validation dataset, the performance of traditional ML and DL models was comparable, and neither class of models demonstrated substantial improvement over catalogue-based approaches, underscoring challenges in cross-dataset generalization. Overall, this benchmarking study provides a comprehensive and systematic evaluation of current approaches, establishes a rigorous evaluation framework for future comparisons, and identifies key methodological considerations necessary to advance robust drug resistance prediction in clinical settings. To enhance reproducibility and facilitate the application of TB-Bench to additional datasets and models, we have made the source code publicly available at https://github.com/BIRDSgroup/TB-Bench.

5
A Machine Learning Approach for Physiological Role Prediction in Protein Contact Networks: a large-scale analysis on the human proteome

Cervellini, M.; Martino, A.

2026-04-14 bioinformatics 10.64898/2026.04.10.717657 medRxiv
Top 2%
0.0%
Show abstract

Proteins are fundamental macromolecules involved in virtually all biological processes. Their physiological roles are tightly linked to their three-dimensional structure, which can be naturally abstracted as Protein Contact Networks (PCNs), i.e., graphs where residues are nodes and edges encode spatial proximity. This representation enables the application of Graph Machine Learning to address the protein functional annotation gap at proteome scale. In this work, protein function prediction is studied on the majority of the human proteome, focusing on enzymatic activity and enzyme class assignment as well-defined and biologically meaningful targets. A large-scale supervised analysis was conducted on PCNs derived from experimentally resolved human protein structures. Multiple graph-based learning paradigms were systematically compared under a unified evaluation protocol, including handcrafted graph embeddings, kernel methods, and end-to-end Graph Neural Networks (GNNs). Feature engineering approaches comprised (i) spectral density embeddings of the normalized graph Laplacian and (ii) higher-order topological representations based on simplicial complexes, with optional INDVAL-based feature selection. These representations were paired with linear, ensemble, and kernel classifiers, while GNNs were trained directly on raw PCNs exploiting a diverse set of message-passing architectures. Two tasks were considered: binary classification of enzymatic versus non-enzymatic proteins and multiclass prediction of first-level Enzyme Commission (EC) classes. Performance was assessed using repeated stratified splits to ensure robust and variance-aware evaluation. In the binary enzymatic classification task, the Jaccard-based graph kernel achieved the best performance with an adjusted balanced accuracy of 0.90, closely followed by GNNs trained end-to-end on PCNs. In the multiclass EC prediction task, GNNs demonstrated superior discriminative power, reaching an adjusted balanced accuracy of 0.92 and outperforming all explicit embedding and kernel-based approaches. Overall, results indicate that EC class prediction is intrinsically more complex than binary enzymatic discrimination and benefits from the higher expressivity of deep message-passing architectures. The findings demonstrate that graph-based representations of protein structure support competitive functional prediction at proteome scale, with classical kernel methods and modern GNNs offering complementary strengths in terms of accuracy, scalability, and flexibility.

6
Time to Discharge and Associated Factors Among Preterm Neonates Admitted to Kiwoko Hospital, Nakaseke District, Uganda: A Competing Risks Analysis

Mutibwa, S.; Wandiembe, S.; Mbonye, K.; Nsimbe, D.

2026-04-15 pediatrics 10.64898/2026.04.13.26350793 medRxiv
Top 2%
0.0%
Show abstract

Background: Preterm births contribute to approximately 35% of neonatal deaths globally, with an estimated 13.4 million infants born prematurely each year. Despite this substantial burden, limited evidence exists on time to discharge and its determinants among preterm neonates admitted to Neonatal Intensive Care Units (NICUs), particularly in rural Ugandan settings. This study aimed to investigate time to discharge and associated factors among preterm neonates admitted to Kiwoko Hospital in Nakaseke District, Uganda. Methods: A retrospective cohort study was conducted using secondary data from Kiwoko Hospital on preterm neonates admitted to the Neonatal Intensive Care Unit (NICU) between 2020 and 2021 (n = 847). The cumulative incidence function was used to estimate the probability of discharge within 28 days of admission, accounting for competing events. A Fine and Gray sub-distribution hazard regression model was fitted to identify factors associated with time to discharge. Results: Of the 847 preterm admissions, 70.1% were discharged alive within 28 days. The median time to discharge was 14 days. The cumulative incidence of discharge by 28 days was 68%, accounting for competing events. During follow-up, 165 neonates did not complete the 28-day period, including 88 deaths. Factors significantly associated with time to discharge included place of delivery (SHR: 0.62; 95% CI: 0.53-0.73; p<0.001), maternal residence in other districts (SHR: 0.69; 95% CI: 0.48-0.99; p=0.044), extreme preterm (SHR: 0.05; 95% CI: 0.03-0.09; p<0.001), very preterm (SHR: 0.18; 95% CI: 0.14-0.25; p<0.001), moderate preterm (SHR: 0.59; 95% CI: 0.46-0.76; p<0.001), triplet births (SHR: 0.40; 95% CI: 0.23-0.68; p=0.001), 2-4 ANC visits (SHR: 0.70; 95% CI: 0.56-0.87; p=0.002), <=1 ANC visit (SHR: 0.64; 95% CI: 0.49-0.85; p=0.002), respiratory distress syndrome (SHR: 0.64; 95% CI: 0.48-0.74; p<0.001), and birth trauma (SHR: 2.62; 95% CI: 1.60-4.29; p<0.001). Conclusions: Respiratory distress syndrome, fewer antenatal care visits, out-of-district residence, and higher degrees of prematurity were associated with prolonged time to discharge among preterm neonates. Strengthening antenatal care utilization and improving access to quality neonatal care in underserved areas may enhance discharge outcomes.

7
Triage Administration of Ondansetron for Gastroenteritis in children; a randomized controlled trial

Weill, O.; Lucas, N.; Bailey, B.; Marquis, C.; Gravel, J.

2026-04-15 pediatrics 10.64898/2026.04.13.26350796 medRxiv
Top 2%
0.0%
Show abstract

Objectives: Acute gastroenteritis is a leading cause of pediatric emergency department (ED) visits. While ondansetron reduces vomiting, intravenous rehydration, and hospital admissions, its efficacy when initiated at triage remains unclear. We aimed to evaluate whether triage nurse-initiated administration of ondansetron in children with suspected gastroenteritis reduces the proportion of patients requiring observation following initial physician assessment. Methods: We conducted a randomized, double-blind, placebo-controlled trial in a tertiary pediatric ED in Canada. Children aged 6 months to 17 years presenting with morae than 3 episodes of vomiting in the preceding 24 hours (including 1 within 2 hours of arrival), were eligible. At triage, we randomized participants to receive liquid ondansetron or a color- and taste-matched placebo. The primary outcome was the proportion of patients requiring observation after the first physician evaluation. Secondary outcomes included post-intervention vomiting, ED length of stay, patient comfort, and 48-hour return visits. The trial was registered at ClinicalTrials.gov (NCT03052361). Results: Recruitment was stopped prematurely due to the COVID-19 pandemic. Ninety-one participants were randomized to ondansetron (n= 44) or placebo (n= 47). Overall, 40 patients (45%) were discharged immediately after the initial physician assessment, with no difference between the ondansetron and placebo groups (44% vs. 45%; absolute difference -1%, 95% CI: -20% to 19%). No significant differences were observed in all secondary outcomes. Conclusion: In this trial, triage nurse-initiated ondansetron administration did not reduce the need for ED observation in children with presumed gastroenteritis. While being underpowered, this study could inform researchers planning larger clinical trials.

8
Understanding response to treatment in depression: Insights from the Pakistani DIVERGE study

Umar, M.; Hussain, F.; Khizar, B.; Khan, I.; Khan, F.; Cotic, M.; Chan, L.; Hussain, A.; Ali, M. N.; Gill, S. A.; Mustafa, A. B.; Dogar, I. A.; Nizami, A. T.; Haq, M. M. u.; Mufti, K.; Ansari, M. A.; Hussain, M. I.; Choudhary, S. T.; Maqsood, N.; Rasool, G.; Ali, H.; Ilyas, M.; Tariq, M.; Shafiq, S.; Khan, A. A.; Rashid, S.; Ahmad, H.; Bettani, K. U.; Khan, M. K.; Choudhary, A. R.; Mehdi, M.; Shakoor, A.; Mehmood, N.; Mufti, A. A.; Bhatia, M. R.; Ali, M.; Khan, M. A.; Alam, N.; Naqvi, S. Q.-i.-H.; Mughal, N.; Ilyas, N.; Channar, P.; Ijaz, P.; Din, A.; Agha, H.; Channa, S.; Ambreen, S.; Rehman,

2026-04-17 psychiatry and clinical psychology 10.64898/2026.04.13.26350625 medRxiv
Top 2%
0.0%
Show abstract

BackgroundMajor depressive disorder (MDD), a leading cause of disability worldwide, exhibits substantial heterogeneity in treatment outcomes. Patients who do not respond to standard antidepressant therapy account for the majority of MDDs disease burden. Risk factors have been implicated in treatment response, including genes impacting on how antidepressants are metabolised. Yet, despite its clinical importance, risk factors for treatment-resistant depression (TRD) remain unexplored in low- and middle-income countries (LMIC). We used data from the DIVERGE study on MDD to investigate the risk factors of TRD in Pakistan. MethodsDIVERGE is a genetic epidemiological study that recruited adult MDD patients ([&ge;]18 years) between Sep 27,2021 to Jun 30, 2025, from psychiatric care facilities across Pakistan. Detailed phenotypic information was collected by trained interviewers and blood samples taken. Infinium Global Diversity Array with Enhanced PGx-8 from Illumina was used for genotyping followed by DRAGEN calling to infer metaboliser phenotypes for Cytochrome P450 (CYP) enzyme genes. We defined TRD as minimal to no improvement after [&ge;]12 weeks of adherent antidepressant therapy. We conducted multi-level logistic regression to test the association of demographic, clinical and pharmacogenetic variables with TRD. FindingsAmong 3,677 eligible patients, polypharmacy was rampant; 86% were prescribed another psychotropic drug along with an antidepressant. Psychological therapies were uncommon (6%) while 49% of patients had previously visited to a religious leader/faith healer in relation to their mental health problems. TRD was experienced by 34% (95%CI: 32-36%) patients. The TRD group was characterised by more psychotic symptoms and suicidal behaviour (OR=1.39, 95%CI=1.04-1.84, p=0.02; OR=1.03, 95%CI=1.01-1.05, p=0.005). Social support (OR=0.55, 95%CI=0.44-0.69, p=1.4x10-7) and parents being first cousins (OR=0.81, 95%CI=0.69-0.96, p=0.01) were associated with lower odds of TRD. In 1,085 patients with CYP enzyme data, poor (OR=1.85, 95%CI=1.11-3.07, p=0.01) and ultra-rapid (OR=3.11, 95%CI=1.59-6.12, p=0.0009) metabolizers for CYP2C19 had increased risk of TRD compared with normal metabolisers. InterpretationThere was an excessive use of polypharmacy in the treatment of depression while psychological therapies were uncommon highlighting the need for more evidence-based practice. This first large study of MDD from Pakistan uncovered the importance of culture-specific forms of social support in preventing TRD, highlighting opportunities for interventions in low-income settings. Pharmacogenetic markers can be leveraged to predict TRD.

9
Clinical and Genetic Evaluation of Suicide Death with and without Interpersonal Trauma Exposure

Monson, E. T.; Shabalin, A. A.; Diblasi, E.; Staley, M. J.; Kaufman, E. A.; Docherty, A. R.; Bakian, A. V.; Coon, H.; Keeshin, B. R.

2026-04-16 psychiatry and clinical psychology 10.64898/2026.04.14.26350901 medRxiv
Top 2%
0.0%
Show abstract

Importance: Suicide is a leading cause of death in the United States with risk strongly influenced by Interpersonal trauma, contributing to treatment resistance and clinical complexity. Objective: To assess clinical and genetic factors in individuals who died from suicide, with and without interpersonal trauma exposure. Design: Individuals who died from suicide with and without trauma were compared in a retrospective case-case design. Prevalence of 19 broad clinical categories was assessed between groups. Results directed selection of 42 clinical subcategories, and 40 polygenic scores (PGS) for further assessment. Multivariable logistic regression models, adjusted for critical covariates and multiple tests, were formulated. Models were also stratified by age group (<26yo and >=26yo), sex, and age/sex. Setting: A population-based evaluation of comorbidity and polygenic scoring in two suicide death subgroups. Participants: A total of 8 738 Utah Suicide Mortality Research Study individuals (23.9% female, average age = 42.6 yo) who died from suicide were evaluated, divided into trauma (N = 1 091) and non-trauma exposed (N = 7 647) individuals. A subset of unrelated European genotyped individuals was also assessed in PGS analyses (Trauma N = 491; Non-trauma N = 3 233). Exposures: Trauma is here defined as interpersonal trauma exposure, including abuse, assault, and neglect from International Classification of Disease coding. Main Outcomes and Measures: Prevalence of comorbid clinical sub/categories and PGS enrichment in trauma versus non-trauma exposed suicide deaths. Results: Overall, trauma-exposed individuals died from suicide earlier (mean age of 38.1 yo versus 43.3 yo; P <0.0001) and were disproportionately female (38% versus 21%, OR = 3.3, CI = 2.9-3.8). Prevalence of asphyxiation and overdose methods, prior suicidality, psychiatric diagnoses, and substance use (OR range = 1.3-3.7) were elevated in trauma exposed individuals who died from suicide. Genetic PGS were also elevated in trauma-exposed individuals who died from suicide for depression, bipolar disorder, cannabis use, PTSD, insomnia, and schizophrenia (OR range = 1.1-1.4) with ADHD and opioid use showing uniquely elevated PGS in trauma exposed males (OR range = 1.2-1.4). Conclusions and Relevance: Results demonstrated multiple convergent lines of age- and sex-specific evidence differentiating trauma-exposed from non-trauma exposed suicide death. Such findings suggest unique biological backgrounds and may refine identification and treatment of this high-risk group.

10
Characteristics of individuals with cerebral palsy across the United States

Aravamuthan, B. R.; Bailes, A. F.; Baird, M.; Bjornson, K.; Bowen, I.; Bowman, A.; Boyer, E.; Gelineau-Morel, R.; Glader, L.; Gross, P.; Hall, S.; Hurvitz, E.; Kruer, M. C.; Larrew, T.; Marupudi, N.; McPhee, P.; Nichols, S.; Noritz, G.; Oleszek, J.; Ramsey, J.; Raskin, J.; Riordan, H.; Rocque, B.; Shah, M.; Shore, B.; Shrader, M. W.; Spence, D.; Stevenson, C.; Thomas, S. P.; Trost, J.; Wisniewski, S.

2026-04-16 pediatrics 10.64898/2026.04.14.26350870 medRxiv
Top 2%
0.0%
Show abstract

Objective Cerebral palsy (CP) affects approximately 1 million Americans and 18 million individuals worldwide, yet contemporary US epidemiologic data remains limited. We aimed to use Cerebral Palsy Research Network (CPRN) clinical registry to describe demographics and clinical characteristics of individuals with CP across the US and determine associations with gross motor function and genetic etiology. Methods Registry subjects were included if they had clinician-confirmed CP and prospectively entered data for Gross Motor Function Classification System (GMFCS) Level, gestational age, genetic etiology, CP distribution, and tone/movement types. Logistic regression was used to determine which of these variables plus race, sex, ethnicity, and age were associated with GMFCS level and genetic etiology. Results A total of 9,756 children and adults with CP from 22 CPRN sites met inclusion criteria. Participants were predominantly White (73.0%), male (57.3%), non-Hispanic (87.8%), and younger than 18 years (73.7%). Most were classified as GMFCS levels I-III (55.6%), born preterm (52.8%), had spasticity (83.8%), and had quadriplegia (41.9%); 12.2% were identified as having a genetic etiology. Tone/movement types, CP distribution, and gestational age were significantly associated with both GMFCS level and genetic etiology (p<0.001). Compared to White individuals, Black individuals were more likely to have greater gross motor impairment (p<0.001). Conclusion In this large US cohort, clinical and demographic factors, including race, were associated with gross motor function and genetic etiology in CP. These findings highlight persistent disparities and demonstrate the value of a national clinical registry for informing prognostication, quality improvement efforts, and targeted genetic testing strategies.

11
An independent supervisory safety agent improves reaction of large language models to suicidal ideation

Trivedi, S.; Simons, N. W.; Tyagi, A.; Ramaswamy, A.; Nadkarni, G. N.; Charney, A. W.

2026-04-15 psychiatry and clinical psychology 10.64898/2026.04.13.26350757 medRxiv
Top 2%
0.0%
Show abstract

Background: Large language models (LLMs) are increasingly used in mental health contexts, yet their detection of suicidal ideation is inconsistent, raising patient safety concerns. Objective: To evaluate whether an independent safety monitoring system improves detection of suicide risk compared with native LLM safeguards. Methods: We conducted a cross-sectional evaluation using 224 paired suicide-related clinical vignettes presented in a single-turn format under two conditions (with and without structured clinical information). Native LLM safeguard responses were compared with an independent supervisory safety architecture with asynchronous monitoring. The primary outcome was detection of suicide risk requiring intervention. Results: The supervisory system detected suicide risk in 205 of 224 evaluations (91.5%) versus 41 of 224 (18.3%) for native LLM safeguards. Among 168 discordant evaluations, 166 favored the supervisory system and 2 favored the LLM (matched odds ratio {approx}83.0). Both systems detected risk in 39 evaluations, and neither in 17. Detection was highest in scenarios with explicit suicidal ideation and lower in more ambiguous presentations. Conclusions: Native LLM safeguards frequently failed to detect suicide risk in this structured evaluation. An independent monitoring approach substantially improved detection, supporting the role of external safety systems in high-risk mental health applications of LLMs.

12
Strategies to enroll and retain low-income adolescent and young adult pregnant women in longitudinal studies: lessons learned from the AMOR project

Camara, S. M. A.; de Souza Barbosa, J. F.; Hipp, S.; Fernandes Macedo, S. G. G.; Sentell, T.; Bassani, D. G.; Domingues, M. R.; Pirkle, C. M.

2026-04-17 public and global health 10.64898/2026.04.13.26350540 medRxiv
Top 2%
0.0%
Show abstract

BackgroundProspective studies of pregnant adolescents are essencial to effectively address this global health priority. They help answer vital questions about their health, but such studies are uncommon due to the difficulty in retaining adolescents. This paper describes the successes and challenges of the research strategies used to ensure sufficient recruitment and retention of pregnant adolescents in a longitudinal study about adolescent childbearing in an under-resourced setting. MethodsThe Adolescence and Motherhood Research project was conducted in a rural region of Northeast Brazil in 2017-2019 and assessed 50 primigravids between 13-18 years (adolescents) and 50 primigravids between 23-28 years (young adults) during the first 16 weeks of pregnancy with two follow-ups (third trimester of pregnancy, and 4-6 weeks postpartum). Recruitment strategies involved engagement of health sector and community, as well as referrals from health care professionals and dissemination of the project in different locations. Retention strategies included maintaining contact with the participants between assessments and providing transportation for them to attend the follow-up procedures. ResultsRecruitment took 10 months to complete. A total of 78% of the participants were recruited from the primary health care units, mainly after referral from a health care provider. Retention reached 95% of the sample throughout the study (90%: adolescents; 98%: adults). ConclusionA combination of approaches is necessary to successfully recruit and retain youth in longitudinal studies and engaging local stakeholders may help to increase community-perceived legitimacy of the research. Working closely with front-line staff is essential when conducting research in rural low-income communities.

13
Adherence in Monitoring of ART response and turnaround time of results as per HIV viral load testing guideline among people living with HIV in Dar es salaam Region.

Masegese, T.; MUNG'ONG'O, G. S.; Kamala, B.; Anaeli, A.; Bago, M.; Mtoro, M. J.

2026-04-16 public and global health 10.64898/2026.04.14.26350908 medRxiv
Top 2%
0.0%
Show abstract

Background: HIV/AIDS remains a major public health challenge in Tanzania, where viral load suppression among adults on ART stands at 78% and HVL testing uptake among eligible patients is approximately 22%. Since the introduction of the National HVL Testing Guideline in 2015, little has been done to systematically evaluate its implementation. Objective: To evaluate adherence to the National HVL Testing Guideline across CTC clinics in Dar es Salaam Region, covering ART monitoring, documentation, turnaround time, and factors affecting implementation. Methods: A cross-sectional study was conducted in 2021 across 15 public health facilities with CTC clinics in all five Dar es Salaam districts. A total of 330 PLHIV on ART for more than six months were selected through systematic random sampling with proportional to size allocation, and 45 healthcare providers through convenient sampling. Data were collected via abstraction forms and self-administered questionnaires, and analysed using SPSS Version 23 with descriptive statistics, bivariate analysis, and binary logistic regression. Results: Only 25.1% of patients had their first HVL sample taken at six months as per guideline, with 68.8% delayed beyond six months. Second and third samples were similarly delayed. MoHCDGEC sample tracking forms were absent in 96.7% of facilities and incomplete in 99.1%, and no facility captured specimen acceptance or rejection as site feedback. Turnaround time exceeded the 14-day guideline threshold in 64.5%, 66.7%, and 69.4% of first, second, and third results respectively. Patient negligence (AOR=9.84; 95% CI: 1.83-52.77) and storage (AOR=5.72; 95% CI: 0.94-35.0) were independently associated with guideline adherence. Conclusion: Adherence to the National HVL Testing Guideline in Dar es Salaam is suboptimal across testing timelines, documentation, and turnaround time, with patient negligence and storage capacity as significant determinants. Targeted interventions are needed to strengthen patient education, improve storage infrastructure, enhance documentation systems, and support providers in adhering to guideline-specified timelines.

14
Female genital cutting and maternal attitudes about it: Testing a cultural disempowerment hypothesis

Strand, P. S.; Trang, J. C.

2026-04-16 public and global health 10.64898/2026.04.14.26350909 medRxiv
Top 2%
0.0%
Show abstract

Female genital cutting (FGC) is identified within global health and human rights discourse as aligned with gender inequality and female disempowerment. The persistence of FGC in high-prevalence societies is assumed to reflect womens limited influence over decisions concerning their daughters. Yet anthropological research has questioned whether this interpretation adequately reflects how FGC is organized within practicing communities. Across two studies with 176,728 participants from 15 African and Asian countries, we examine whether mothers attitudes toward FGC predict daughters circumcision status and whether this relationship varies with regional FGC prevalence. Multilevel logistic regression models show that maternal attitudes strongly predict daughter circumcision status across both datasets. Contrary to expectations derived from disempowerment frameworks, the association between maternal attitudes and daughter outcomes is not weaker in high-prevalence contexts, it is stronger. These findings suggest that interpretations of FGC as reflecting female disempowerment may mischaracterize the social dynamics of societies in which FGC is common. Policy implications of the findings are discussed.

15
Baseline Assessment of Drug-Drug Interaction Knowledge Among Healthcare Providers in Kibaha, Tanzania

Salim, A.; Allen, M.; Mariki, K.; Pallangyo, T.; Maina, R.; Mzee, F.; Minja, M.; Msovela, K.; Liana, J.

2026-04-16 public and global health 10.64898/2026.04.11.26350082 medRxiv
Top 2%
0.0%
Show abstract

In the context of global health, the ability of frontline primary health providers to identify potential Drug-Drug Interactions (DDIs) is a critical component of patient safety. This is particularly true in settings like Tanzania, where drug dispensers often serve as the primary point of contact for healthcare. In this study, we establish a baseline for drug decision-making capabilities across multiple cadres of healthcare providers in Kibaha, Tanzania. We specifically distinguish between the ability to recognize safe drug combinations versus harmful ones. The findings reveal a critical asymmetry in provider performance: while professional training improves the recognition of safe combinations, it provides no advantage over lay intuition (and in some cases, a significant disadvantage) in detecting potentially harmful interactions.

16
Accumulation of Benzalkonium Chloride from Disinfectants in Dust Associated with Increased Microbial Tolerance

Yu, J.; Tillema, S.; Akel, M.; Aron, A.; Espinosa, E.; Fisher, S. A.; Branche, T. N.; Mithal, L. B.; Hartmann, E. M.

2026-04-16 public and global health 10.64898/2026.04.14.26350823 medRxiv
Top 2%
0.0%
Show abstract

Benzalkonium chloride (BAC) is widely used as a disinfectant in cleaning products and is frequently detected in indoor dust. In this study, we assessed dust samples, along with information on cleaning product use, from 24 pregnant participants. Dust samples were analyzed for BAC concentration and microbial tolerance. Different chain lengths of BAC (C12, C14, and C16) were quantified using LC-MS/MS, and bacterial isolates were tested for BAC tolerance using minimum inhibitory concentration (MIC) assays. BAC was ubiquitously detected, with C12 and C14 being dominant. Higher BAC concentrations were associated with reported disinfectant use and increased microbial tolerance. These findings suggest that indoor antimicrobial use may promote microbial resistance, highlighting potential exposure risks in indoor environments and the need for further investigation into health and ecological impacts.

17
Strengthening School Water, Sanitation and Hygiene (WASH) Programme Implementation: Evidence from Expert Consensus in Uasin Gishu County, Kenya

SERONEY, G. C.; Magak, N. A. G.; Mchunu, G. G.

2026-04-16 public and global health 10.64898/2026.04.14.26350916 medRxiv
Top 2%
0.0%
Show abstract

Introduction Access to safe water, sanitation, and hygiene (WASH) in schools is critical for child health, learning, and gender equity. In Kenya, the Kenya School Health Policy and the Basic Education Act outline standards for school WASH; however, implementation remains uneven due to inadequate infrastructure, weak inter-sectoral coordination, and limited financing. This study aimed to identify priority components for strengthening school WASH implementation and generate policy-relevant recommendations based on expert consensus in Uasin Gishu County, Kenya. Methods and Results A Delphi technique consisting of two iterative rounds was used to reach expert consensus. In Round 1, 20 purposively selected experts including head teachers, county education officials, public health officers, water and public works officers, and NGO representatives participated in key informant interviews. Emergent themes informed development of a structured Round 2 questionnaire administered through CommCare online app. Quantitative data were analyzed using descriptive statistics (means, standard deviations, percentage agreement), while qualitative responses underwent thematic coding using NVivo 12. Experts reached strong consensus on essential components required for strengthening school WASH implementation. Key priorities included clear governance structures, designated budget lines, inclusive infrastructure, menstrual hygiene management (MHM), curriculum integration, sustained capacity building, and systematic monitoring. Multi-sectoral collaboration and recognition of best-performing schools were also emphasized as important motivators for compliance and sustainability. Equity considerations particularly the need for disability-friendly facilities and school-community outreach were highlighted as critical. Agreement levels ranged from 74% to 100%, with most items scoring mean values between 4.5 and 4.8 on a 5-point Likert scale, indicating strong consensus among experts. Conclusion strengthening implementation of school WASH in Kenya requires coordinated governance, predictable funding, reliable water systems, inclusive sanitation, strengthened MHM, and consistent monitoring beyond infrastructure investment alone. Integrating these expert-validated priorities within existing national policies offers a practical pathway to improving learner health, reducing absenteeism especially among girls and promoting equitable educational outcomes.

18
Evaluating Large Language Models for Transparent Quality-of-Care Measurement in Children with ADHD

Bannett, Y.; Pillai, M.; Huang, T.; Luo, I.; Gunturkun, F.; Hernandez-Boussard, T.

2026-04-17 pediatrics 10.64898/2026.04.12.26350732 medRxiv
Top 2%
0.0%
Show abstract

ImportanceGuideline-concordant care for young children with attention-deficit/hyperactivity disorder (ADHD) includes recommending parent training in behavior management (PTBM) as first-line treatment. However, assessing guideline adherence through manual chart review is time-consuming and costly, limiting scalable and timely quality-of-care measurement. ObjectiveTo evaluate the accuracy and explainability of large language models (LLMs) in identifying PTBM recommendations in pediatric electronic health record (EHR) notes as a scalable alternative to manual chart review. Design, Setting, and ParticipantsThis retrospective cohort study was conducted in a community-based pediatric healthcare network in California consisting of 27 primary care clinics. The study cohort included children aged 4-6 years with [&ge;] 2 primary care visits between 2020-2024 and ICD-10 diagnoses of ADHD or ADHD symptoms (n=542 patients). Clinical notes from the first ADHD-related visit were included. A stratified subset of 122 notes, including all cases with model disagreement, was manually annotated to assess model performance in identifying PTBM recommendations and rank model explanations. ExposuresAssessment and plan sections of clinical notes were analyzed using three generative large language models (Claude-3.5, GPT-4o, and LLaMA-3.3-70B) to identify the presence of PTBM recommendations and generate explanatory rationales and documentation evidence. Main Outcomes and MeasuresModel performance in identifying PTBM recommendations (measured by sensitivity, positive predictive value (PPV), and F1-score) and qualitative explainability ratings of model-generated rationales (based on the QUEST framework). ResultsAll three models demonstrated high performance compared to expert chart review. Claude-3.5 showed balanced performance (sensitivity=0.89, PPV=0.95, and F1-score=0.92) and ranked highest in explainability. LLaMA3.3-70B achieved sensitivity=0.91, PPV=0.89, and F1-score=0.90, ranking second for explainability. GPT-4o had the highest PPV [0.97] but lowest sensitivity [0.82], with an F1-score of 0.89 and the lowest explainability ranking. Based on classifications from the best-performing model, Claude-3.5, 26.4% (143/542) of patients had documented PTBM recommendations at their first ADHD-related visit. Conclusions and RelevanceLLMs can accurately extract guideline-concordant clinician recommendations for non-pharmacological ADHD treatment from unstructured clinical notes while providing clear explanations and supporting evidence. Evaluating model explainability as part of LLM implementation for medical chart review tasks can promote transparent and scalable solutions for quality-of-care measurement.

19
Clinical Application of CT-Guided Lung Nodule Localization Needles in Preoperative Localization of Small Pulmonary Nodules

Xu, R.; Dou, H.; Zhang, M.; Liu, Z.

2026-04-16 surgery 10.64898/2026.04.13.26350830 medRxiv
Top 2%
0.0%
Show abstract

Background: To investigate the safety and efficacy of CTguided lung nodule localization needles for the preoperative localization of small pulmonary nodules. Methods: A retrospective study was conducted on 102 patients with a total of 113 small pulmonary nodules who underwent preoperative localization at Jinan Fourth People's Hospital from January 2024 to December 2025. Nodule diameter and depth, localization time, the number of pleural punctures, the localization success rate, and postoperative complications (hook dislodgement, hemorrhage, and pneumothorax) were recorded. All patients underwent video assisted thoracoscopic surgery (VATS) after localization. Results: The mean nodule diameter was 0.97{+/-}0.36 cm, the mean depth was 1.26{+/-}0.48 cm, and the mean localization time was 9.8{+/-}3.65 minutes. The hook dislodgement rate was 0.98% (1/102), the intrapulmonary hemorrhage rate was 14.71% (15/102), and the pneumothorax rate was 16.67% (17/102). All pulmonary nodules were successfully resected by VATS at 73.82{+/-}13.83 minutes after localization, and no severe complications occurred. Conclusions: The use of a CTguided lung nodule localization needle for the preoperative localization of small pulmonary nodules decreases the time needed for intraoperative nodule detection and operation time. This strategy is a simple, safe, and accurate preoperative localization method that is worthy of increased clinical use.

20
Variation in Haemostasis and VTE Prophylaxis in Elective Adult Cranial Neurosurgery: A Global Survey of Perioperative Practice

Pandit, A. S.; Chaudri, T.; Chaudri, Z.; Vasilica, A. M.; Dhaliwal, J.; Sayar, Z.; Cohen, H.; Westwood, J. P.; Toma, A. K.

2026-04-16 surgery 10.64898/2026.04.14.26350905 medRxiv
Top 2%
0.0%
Show abstract

Background Venous thromboembolism (VTE) remains a major cause of perioperative morbidity in cranial neurosurgery, yet clinical practice varies widely, and formal guidelines are inconsistent. Understanding internationally sampled neurosurgical practice is essential for informing consensus and future trials. Methods An international, 2-stage cross-sectional, internet-based survey was conducted. Practising neurosurgeons performing elective adult cranial surgery were eligible. Descriptive statistics were used to summarise practice. Responses covered patterns of pre-operative haemostasis decision making, use and timing of mechanical and/or chemical prophylaxis, use of perioperative imaging prior to anticoagulation, and frequency of clinical assessment for VTE. Associations with geographical income status, subspecialty, and years post-certification were statistically tested. Practice heterogeneity was quantified and contextual influence was summarised using mean effect sizes across stratifying variables in order to determine domains of true equipoise. Results Of 585 responses, 456 (78%) met criteria for inclusion: representing 322 units across 78 countries (71% high-income). Thirteen per cent reported no departmental VTE plan; 23% followed no guidelines and 12% used multiple. Routine pre-operative testing almost universally included haemoglobin/platelets/haematocrit, with fibrinogen more common in high-income settings. Compared with high-income country respondents, low- and middle-income respondents reported higher haemoglobin transfusion thresholds (>90 g/dL; p<0.001) and shorter antiplatelet interruption (p[&le;]0.03), and less frequent outpatient VTE assessment (p<0.001). Mechanical prophylaxis was common (TEDs 81%, IPC 62%), typically started pre- or intra-operatively. Among those completing the chemoprophylaxis section (n=310), 57% required a CT or MRI scan before LMWH which was then initiated on average 31.4 hours after surgery. 1% of respondents did not routinely use LMWH. Many clinical decisions demonstrated statistical equipoise ie. high heterogeneity with low contextual influence. Conclusion Peri-operative haemostasis and VTE prophylaxis practices in adult elective cranial neurosurgery vary substantially worldwide, with some decisions reflecting geographical or socioeconomic differences and many others reflecting true clinical equipoise rather than contextual determinants. By mapping contemporary real-world practice across diverse health-system contexts, this study provides a necessary empirical foundation for rational trial design and future guideline development.