SoftwareX
○ Elsevier BV
Preprints posted in the last 90 days, ranked by how well they match SoftwareX's content profile, based on 15 papers previously published here. The average preprint has a 0.02% match score for this journal, so anything above that is already an above-average fit.
Kern, N. R.; Park, S.; Cao, Y.; Im, W.
Show abstract
As high-performance computing provides the ability to generate and analyze ever larger simulation trajectories, the challenges in learning, applying, and sharing the best analytical practices become more salient. Extracting reproducible scientific insights from simulation requires a thorough understanding of many computing topics unrelated to the molecular systems being modeled and simulated. While the rapid development of the technologies used for analysis makes previously impossible studies into routine work, the growing repertoire of software combined with the specificity of the ecosystems that they rely on can easily break the programs used in older studies. In this work, we present ST-Analyzer, a simulation trajectory analysis suite with command-line (CLI) and graphical (GUI) user interfaces. ST-Analyzer is distributed freely as an open-source conda-forge package with support for macOS, Linux, and Windows (via WSL2). Besides facilitating several common analysis tasks, the GUI shows users the exact commands necessary to repeat the same tasks on the command-line. We demonstrate ST-Analyzers capabilities by reproducing several results from previously published simulation studies on the lipid parameters of heterogeneous biomembranes and the behavior of a SARS-CoV-2 spike protein-antibody complex. We expect ST-Analyzer to be useful to experts for quickly setting up common analysis tasks and to nonexperts as a guided introduction to simulation analysis using both GUI and CLI. ST-Analyzer is freely available at https://github.com/nk53/stanalyzer.
Bright, M.; Mi, X.; Duarte, D.; Carey, E.; Lyu, B.; Wang, Y.; Nimmerjahn, A.; Yu, G.
Show abstract
BackgroundAdvanced biological imaging analysis platforms such as Activity Quantification and Analysis (AQuA2) enable accurate spatiotemporal activity analysis across diverse cell populations within many species. These tools are increasingly important for investigating cellular signaling dynamics and behavior. However, despite advances in the accuracy and species capability of AQuA2, it remains computationally demanding for analysis of long time-series datasets and requires all users to maintain a MATLAB license, which may limit accessibility and large-scale deployment. ResultsTo address these limitations, we have designed and made available AQuA2-Cloud, a portable software stack and web platform developed as an improvement and further evolution of AQuA2. This container-deployable system permits multi-user cloud-based high accuracy activity quantification with intuitive workflows, export of analysis data and project files, and comparable processing times. The platform offers integrated features such as in-browser analysis control interfaces, asynchronous program state control, multiple users and user management, support for unreliable connections, file uploading and downloading via web browsers and File Transfer Protocol, and centralized organization of analysis output. ConclusionAQuA2-Cloud constitutes a cloud-native solution for laboratories or research groups seeking to centralize analysis of spatiotemporal biological imaging datasets while reducing software installation and licensing barriers for end users. The platform enables researchers with minimal technical expertise to perform advanced bioimaging analysis through standard web browsers while maintaining the analytical capabilities of AQuA2. AQuA2-Cloud source code, deployment procedures, and documentation are freely available at (https://github.com/yu-lab-vt/AQuA2-Cloud).
Minasandra, P.; Sridhar, V. H.; Roche, D. G.; Planas-Sitja, I.
Show abstract
Real-time tracking and automated response systems are essential for standardising experiments, reducing observer bias, and improving reproducibility in studies of movement and behaviour. However, existing solutions face significant challenges: AI-based tracking systems require expensive hardware and impose computational delays, creating challenges for closed-loop experiments; existing real-time tracking tools lack standardised implementations for response delivery; and steep learning curves limit accessibility for users without programming or computer vision expertise. Here, we introduce TracktorLive, an open-source Python package designed to overcome these challenges through concurrency and a modular, cassette-based architecture. TracktorLive leverages traditional computer vision techniques to perform image-based object detection without the need for expensive hardware or deep learning. By parallelizing object tracking and response delivery into separate, concurrent server and client processes, the software minimizes frame processing time, enabling rapid, real-time analysis and response delivery. User-friendly cassettes--portable code snippets that can be copy-pasted into scripts--enable users with minimal programming experience to implement complex workflows for use in experiments and practical applications. We demonstrate TracktorLives utility through several applications, including microcontroller-based stimulus delivery for location-dependent manipulations; conditional video recording that activates only during events of interest; kinematic-based response triggering using real-time velocity computations; and multi-cassette experimental designs combining multiple functionalities. Detailed tutorials are provided to familiarize users with TracktorLives operation and functionality, and a growing library of cassettes supports diverse applications out of the box. We validated the software by comparing its response timing to human experimenters in a stimulus delivery task involving two fish species, where TracktorLive demonstrated consistently higher accuracy and lower variability, particularly for fast-moving subjects. Beyond experimental biology, TracktorLives unique architecture and versatility could support many different applications in fields ranging from neuroscience to wildlife management. As an open-source software combining accessibility, modularity, and computational efficiency, TracktorLive can help democratize real-time tracking and automated response systems across disciplines.
Mesbah, I.; Klaus, C.; Sotomayor, M.; Sumbul, F.; Rico, F.
Show abstract
Molecular dynamics simulation is a powerful computational technique used for predicting and understanding the dynamic behavior of biomolecular systems. Steered molecular dynamics (SMD) simulations enable the study of force-induced processes in biomolecules, effectively mimicking single-molecule force spectroscopy experiments probing protein unfolding and receptor-ligand unbinding. Given the stochastic nature of these mechanical events, accurately exploring the dynamic behavior of biomolecules and extracting accurate physical information requires several in-silico experiments. This includes performing many pulling simulations at different velocities or force loading rates. The large amount of data obtained from these simulation sets requires efficient automated data processing tools. We present PySteMoDA, a novel Python package with a user-friendly graphical interface specifically designed for constant-velocity SMD data analysis. The automated force peak detection methods reduce user bias, improve accuracy, and accelerate data analysis. The package also allows identification of residues involved in mechanical events through computation of the time-dependent mechanical work and correlation factors between residue pairs. This package not only addresses automated data processing in SMD simulations and accurate parameter extraction, but also significantly enhances accessibility and usability. Through PySteMoDA, users can efficiently analyze simulation data without the barrier of coding, facilitating a wider range of investigations and insights in the field of computational biochemistry and biophysics.
Soares, G. C. d. F.; Varella, A. L. N.; Facundo, H. T.
Show abstract
Oxidative stress results from excessive accumulation of reactive oxygen species (ROS) and plays a central role in numerous physiological and pathological processes. Accurate quantification of antioxidant enzyme activities is therefore essential in redox biology research. However, data analysis for commonly used assays, such as superoxide dismutase (SOD), catalase (CAT), and glutathione peroxidase (GPx), is frequently performed using spreadsheets or manual calculations, which are time-consuming and prone to error. Here, we present Redoxyme, a free, open-source, Python-based graphical user interface designed to standardize and automate the calculation of antioxidant enzyme activities. The software integrates protein normalization, enzyme-specific calculation routines, data visualization, and Excel export within an intuitive interface that does not require programming expertise. Redoxyme was validated using experimental data obtained from animal tissues (rats and mice), demonstrating excellent agreement with manual calculations and established analytical methods. Redoxyme provides a practical solution for improving reproducibility and efficiency in antioxidant enzyme activity analysis. The software is currently distributed as a standalone executable for Windows (locally installed), and an interactive web-based calculator implemented in Streamlit, enabling direct use without local installation. The source code and version-controlled development history are openly accessible via GitHub, promoting transparency, reproducibility, community-driven improvements, and can, in principle, be adapted for other operating systems. Graphical Abstract O_FIG O_LINKSMALLFIG WIDTH=200 HEIGHT=63 SRC="FIGDIR/small/703993v2_ufig1.gif" ALT="Figure 1"> View larger version (10K): org.highwire.dtl.DTLVardef@120cc68org.highwire.dtl.DTLVardef@4be246org.highwire.dtl.DTLVardef@1f47134org.highwire.dtl.DTLVardef@1341100_HPS_FORMAT_FIGEXP M_FIG C_FIG
Terra, R.; Carvalho, D.; Machado, D. J.; Osthoff, C.; Ocana, K.
Show abstract
Advances in High-Performance Computing (HPC) have enabled increasingly complex genomic analyses, including those in phylogenomics. These analyses contribute to understanding the evolution of viruses and pathogens, improving our knowledge of disease transmission, and supporting targeted public health strategies. However, due to the increasing number of tools and processing steps involved, executing these analyses manually, step by step, becomes error-prone and inefficient. To address this challenge, we present HP2NET, a robust framework for reproducible, efficient, and scalable phylogenetic network analysis. HP2NET integrates five workflows based on state-of-the-art tools such as PhyloNetworks and PhyloNet, allowing the analysis of multiple datasets and workflows in a single execution. The framework includes features such as task packaging and data reuse to improve performance and resource utilization in HPC environments. We perform a comprehensive performance evaluation of the software used within HP2NET, identifying bottlenecks and analyzing gains from parallel processing. Data reuse provided up to 15.35% time reduction, for a small dataset, in our experimental environment, while parallel execution of the five pipelines reduced total runtime by up to 90.96% compared to sequential runs. Finally, we validate HP2NET in a real-world case study by analyzing Dengue virus genomes, demonstrating its applicability value for large-scale phylogenetic analyses.
Banerjee, T.; Abubaker-Sharif, B.; Devreotes, P. N.; Iglesias, P. A.
Show abstract
SummaryThe plasma membrane and accompanying cortex serve as one of the major hubs of the signal transduction and cytoskeletal activities that collectively regulate numerous cell physiological processes such as migration, polarity, macropinocytosis, phagocytosis, cytokinesis, etc. Yet, dynamically tracking membrane-cortex associated protein or lipid kinetics over time from live-cell image series remains a challenging task, primarily due to the difficulty of accurately extracting and aligning the cell boundary between consecutive frames, as the cell continuously deforms and moves. Here, we present Membrane Kymograph Generator, a cross-platform software that accepts multichannel time-lapse live-cell fluorescent imaging datasets as input and automates the cumbersome, heuristic process of boundary tracking, inter-frame alignment, and intensity sampling along the boundary. The software implements a rotational offset minimization algorithm that circularly aligns boundaries across consecutive frames by exhaustively searching for the optimal angular shift that minimizes point-to-point distances, while handling variations in boundary point counts due to cell shape changes. The software outputs kymographs that represent spatiotemporal dynamics of different membrane-associated proteins or biosensors, allows users to fine-tune visualization parameters through an interactive interface, and provides built-in correlation analysis tools for multi-channel datasets. Furthermore, the software allows advanced programmatic usage for batch processing and further analysis via a native API. Our validation tests demonstrated that the Membrane Kymograph Generator can be used to accurately track, visualize, and quantitate the spatial kinetics of a wide array of membrane proteins and lipid biosensors over extended time periods, in a variety of cell types, including Dictyostelium amoeba, human neutrophils, mouse macrophages, and different mammalian cancer cells. The GUI-based software is user-friendly, does not require any technical expertise from users, and significantly reduces the manual effort and time required for kymograph generation and downstream analysis, while ensuring high accuracy and reproducibility. Availability and ImplementationMembrane Kymograph Generator is a free and open-source software, licensed under GNU General Public License 3.0 or later. This software is cross-platform: It can be graphically installed on both x86-64 and AArch64/ARM64 computers, running either Windows, macOS, or any standard Linux distribution. The software is distributed as single installer files (and portable executables) targeting specific hardware architectures and operating systems, and hence, it can be installed natively without any dependency resolution. The source code, detailed documentation, specific installers, portable binaries, and test data are freely available at https://github.com/tatsatb/membrane-kymograph-generator. Additionally, since the software is written in Python, it can also be installed inside any Python environment using PIP package manager (package ID: https://pypi.org/project/membrane-kymograph) and can also be interacted via a built-in Python API.
Ochi, S.; Azuma, M.; Hara, I.; Inada, H.; Takabayashi, K.; Osumi, N.
Show abstract
BackgroundLong-term home-cage monitoring is essential to quantify spontaneous locomotor and social behaviors in group-housed mice, but analysis of high-density RFID tracking data remains a barrier to reproducibility. New methodsWe developed IntelliProfiler 2.0, a fully R-based pipeline tailored to the eeeHive 2D floor-mounted RFID array. The workflow performs data import from text logs, preprocessing, coordinate reconstruction, missing-value handling, feature extraction, statistical testing, and visualization in a single environment. Behavioral metrics include travel distance, close contact ratio (CCR), and a newly implemented inter-individual distance metric. ResultsIn four-day recordings of group-housed C57BL/6J mice (8 males and 8 females), IntelliProfiler 2.0 captured circadian phase-dependent locomotion and proximity patterns and reproduced sex-dependent differences consistent with prior analyses while incorporating updated hardware specifications. Radar-chart summaries enabled intuitive comparison of multidimensional behavioral profiles and inter-individual variability across light/dark phases. Comparison with existing methodsCompared with IntelliProfiler 1.0 and multi-tool workflows, IntelliProfiler 2.0 consolidates analysis into a single, script-based R pipeline, reducing operational complexity and improving reproducibility. The updated implementation supports recent manufacturer-driven changes, including antenna renumbering and multi-USB data export. ConclusionsIntelliProfiler 2.0 provides a reproducible, extensible framework for high-throughput behavioral phenotyping of group-housed mice and is scalable across hardware configurations, including simplified single-board recordings. HighlightsO_LIEnd-to-end R pipeline for eeeHive 2D floor-based RFID tracking analysis C_LIO_LIStandardized setup with comprehensive manuals and protocols C_LIO_LIInter-individual distance metric to quantify group spatial structure C_LIO_LICircadian- and sex-dependent behavioral profiling in group-housed mice C_LIO_LIRadar-charts summarize multidimensional behavioral profiles and variability C_LI
Diaz Ochoa, J. G.; Puskaric, M.; Layer, N.; Jensch, A.; Knott, M.; Krohn, A.
Show abstract
Graph-based methods for data representation and analysis are well suited for encoding both data points and their interrelationships. This approach integrates data and topology, enabling the representation of interrelated information. In this study, we represent patient cohorts as cohort graphs and discuss their application for real-world patient data. We particularly focus on developing methods to cluster patients with similar symptoms and examine how bias parameters (such as sex and age group) influence interlinking within CGs, thereby improving results for accurate patient stratification and personalized decision-making in a clinical context. In particular we illustrate how by considering sex and age groups we can improve the symptom-clustering of a patient population with lung and gastro-intestinal cancer. Finally, we discuss the essential role of high-performance computing (HPC) in upscaling analytical methods for CGs.
Pohar, C.; Rekik, Y.; Phan, M. S.; Gallet, B.; Desroches-Castane, A.; Chevallet, M.; Tinevez, J.-Y.; Tillet, E.; Vigano, N.; Jouneau, P.-H.; Deniaud, A.
Show abstract
The liver has a complex architecture composed of millions of lobules. Within these lobules, hepatocytes, the main hepatic cells, are organized in rows separated by blood capillaries known as sinusoids. These capillaries are lined by liver sinusoidal endothelial cells (LSEC) that form a very specific fenestrated endothelium essential for the exchange of metabolites and proteins between the blood and hepatocytes. Alterations in the size and number of LSEC fenestrations are associated with the onset and the progression of various liver diseases. The analysis of liver architecture is thus of utmost importance for advancing our knowledge of liver ultrastructure and its alterations. Liver architecture has been studied since decades, mainly using 2D electron microscopy, and more recently using advanced super-resolution fluorescence microscopy. In recent years, volume electron microscopy techniques, including focused ion beam-scanning electron microscopy (FIB-SEM) progressed and nowadays enable the 3D reconstruction of biological ultrastructures down to nanometer resolution. However, the analysis of large volumes (e.g., several tens of {micro}m3) remains challenging due to various constraints in the segmentation of large datasets. In the current study, we developed a workflow to semi-automatically segment hepatic sinusoids from FIB-SEM mice liver datasets using the CNN-based (convolutional neural network) tool known as "nnU-Net", after fine-tuning a ground truth model. We also implemented tools for semi-automatic quantification of LSEC fenestrae diameters and sinusoid porosity from segmented datasets. This workflow enabled us to compare the distribution of LSEC fenestrae diameters in wild-type versus Bmp9-deleted mice, a hepatic factor known to be involved in fenestration maintenance. Our results confirm the importance of BMP9 for LSEC differentiation. Therefore, the developed methodology represents a valuable tool for characterizing the fenestrated endothelium under various physiological and pathological conditions.
Virag, D.; Virag, A.-M.; Homolak, J.; Kahnau, P.; Babic Perhoc, A.; Krsnik, A.; Mihalic, L.; Knezovic, A.; Osmanovi{acute} Barilar, J.; Cifrek, M.; Trkulja, V.; Salkovic-Petrisic, M.
Show abstract
Home cage monitoring (HCM) captures longitudinal animal behavioural data without human intervention. However, the systems complexity is rarely addressed in their design, increasing the risk of data loss, which wastes workhours, resources, and animal lives. To assess the feasibility of implementing modern, robust architectures in complex operant HCM paradigms, the VersatiLe Autonomous DevIce for Scheduled Learning Assessment Via Wi-Fi (VLADISLAV) was developed and employed to test cognitive deficits in the intracerebroventricular streptozotocin-induced rat model of sporadic Alzheimers disease (sAD). Reliability was modelled against a system architecture common in commercial HCM systems by modelling the failure rate of the devices critical components across typical durations of animal experiments. VLADISLAV assessed multiple cognitive dimensions of a rat model of sAD with automated, scheduled testing. Its design enabled simultaneous, redundant recording to multiple devices in real time, as well as batch remote control and supervision of tens of VLADISLAVs. VLADISLAV is estimated to reduce component failure rate [~]200-fold at {euro}40/device. Data loss due to system failure shouldnt be accepted as a normal occurrence and robust system design is an ethical imperative. VLADISLAVs robustness and utility demonstrate the potential of embedded networked systems, used in other industries and consumer electronics for over a decade. Today, the open source ecosystem enables cost-effective implementation of such architectures in HCM by biomedical researchers with no electronic engineering education, preventing data loss and facilitating researchers and technicians day-to-day work. Considering these findings, it is apparent that the implementation of modern architectures in HCM is long overdue.
sharma, S.; Kumar, S.; Brull, J. B.; Deepika, D.; Kumar, V.
Show abstract
Transcriptomic analysis is considered a powerful approach for biomarker discovery, however still exploring large scale omics dataset to extract meaningful biological insights remains a challenge for biologists. To address this gap, we present ARACRA a fully automated RNA-seq analysis pipeline including entire transcriptomics workflow from raw FASTQ files to the transcriptomics Point of Departure (tPoD) with human-in-the-loop review process. Overall, the analysis is performed in two phases: Phase 1 carries out the acquisition of raw reads, pre-alignment quality control, alignment to reference genome and quantification of gene expression. Whereas, Phase 2 performs statistical analysis including Differential Gene Expression analysis and Dose-Response modelling. Two phases are separated by an extensive quality control step which allows the user to visually inspect the quality of data processed and helps in filtering noise and outlier samples. ARACRA facilitates end-to-end analysis of RNA-Seq data through an interactive web-based application developed on nextflow and streamlit for minimizing computational complexities while ensuring correct downstream processing. Availability and implementationARACRA is freely available online at the GitHub with MIT License and stream lit-based web application: ARACRA. Researchers can use the demo data or even upload their own data to do the analysis. O_FIG O_LINKSMALLFIG WIDTH=200 HEIGHT=78 SRC="FIGDIR/small/716912v1_fig1.gif" ALT="Figure 1"> View larger version (27K): org.highwire.dtl.DTLVardef@15170a9org.highwire.dtl.DTLVardef@1bb9822org.highwire.dtl.DTLVardef@1010f3aorg.highwire.dtl.DTLVardef@8ee6e6_HPS_FORMAT_FIGEXP M_FIG O_FLOATNOFig 1:C_FLOATNO Overall Architecture of ARACRA C_FIG
Lu, Y.; Pan, M.; Jamwal, V.; Locop, J.; Ruparelia, A. A.; Currie, P. D.
Show abstract
Quantitative histological analysis of skeletal muscle morphometry provides critical insights into muscle physiology but remains labor-intensive and technically demanding. While recent developments in machine-learning-based image segmentation techniques have facilitated large-scale tissue analysis, existing tools that automate muscle morphometry analysis are largely tailored to mammalian models, with limited applicability to teleosts. Moreover, there is a lack of effective tools for visualizing spatial organization and morphometric variability of teleost muscle fibers, a feature that is important for understanding hyperplastic muscle growth dynamics in teleosts. In this study, we show that cytoplasmic staining combined with deep learning-based cell segmentation offers a robust and accurate approach for automated muscle morphometry analysis in developing zebrafish. We also introduce a FIJI2 plugin, implemented in Jython, that streamlines both morphometric analysis and visualization. This tool accommodates shallow and deep learning-based segmentation techniques and incorporates novel quantification and visualization methods suited to teleost-specific muscle features, including mosaic hyperplasia dynamics. The plugin features an intuitive graphical user interface and is designed for flexibility, with minimal constraints regarding species, image quality, or staining protocol. Its modular architecture allows it to be used as a baseline for automated muscle morphometry analysis, while permitting integration with other tools and workflows.
Melykuti, B.; Bustos-Quevedo, G.; Prinz, T.; Nazarenko, I.
Show abstract
Accurate and transparent characterization of extracellular vesicle (EV) preparations is essential to ensure reproducibility, comparability, and adherence to MISEV reporting standards. However, data outputs from commonly used instruments for assessing EV size, concentration, and surface charge (zeta potential) vary widely in format and structure, complicating standardized analysis and integration across platforms. We present PHoNUPS (Plotting the Histogram of Non-Uniform Particles Sizes), free and open-source software (FOSS) developed in R, that enables unified processing, analysis, and visualization of EV characterization data. PHoNUPS computes statistics and generates standardized histograms and contour plots (for size against zeta potential) suitable for transparent reporting and cross-study comparison. The software produces high-quality, publication-ready figures. Third-party graphical editing tools allow users to refine and annotate visualizations for presentation or manuscript preparation. PHoNUPS supports multiple measurement file formats, thereby facilitating dataset integration from different instruments. PHoNUPS was developed with extensibility at its core, providing a basis for user-driven growth. We invite the EV community--researchers, analysts, and tool developers--to use PHoNUPS, share feedback on their experience and needs, and contribute to the platform by integrating additional input data formats, analytical routines, and visualization functionalities. Graphical abstractThe free software PHoNUPS processes the outputs of several different EV characterization instruments and it is extensible with further ones. It computes statistics of particle size and zeta potential distributions and it plots the corresponding histograms or contour plots. O_FIG O_LINKSMALLFIG WIDTH=200 HEIGHT=146 SRC="FIGDIR/small/702479v1_ufig1.gif" ALT="Figure 1"> View larger version (28K): org.highwire.dtl.DTLVardef@b2b3a1org.highwire.dtl.DTLVardef@2f2907org.highwire.dtl.DTLVardef@2ec521org.highwire.dtl.DTLVardef@903624_HPS_FORMAT_FIGEXP M_FIG C_FIG
Cummings, C. E.; Bastien, B. L.; Martinez, J. A.; Luo, J.; Thyme, S. B.
Show abstract
Quantitative phenotyping is essential to studies of animal behavior, enabling systematic analysis of variation arising from natural diversity or experimental manipulation. High-throughput behavioral assays that can simultaneously test multiple animals support sufficiently powered studies of behavioral variation, but accurate tracking of each animal is critical. Furthermore, behavioral tasks and experimental arenas span a wide range of complexity, from the reaction of a single larval zebrafish to an acoustic stimulus to associative conditioning in cue-rich environments. Here, we developed and validated StrIPETrack (Structural similarity-based Image Processing for Estimation and Tracking), a Python-based, modular animal tracking software designed for flexible region-of-interest (ROI) definitions and extensibility across assays. We show that StrIPETrack measures activity comparably to our previous LabVIEW-based zebrafish tracking software and detects similar behavioral differences between wild-type clutches. In addition, StrIPETrack accurately captures behavior in a complex arena: the Y-maze. Our approach for analyzing Y-maze navigation yields an expanded set of metrics beyond turn count and direction, revealing more subtle behavioral variation. Overall, this versatile software can be applied to monitor the activity of multiple animals in parallel in both simple high-throughput and more complex assays, and can be readily adapted to new paradigms. SummaryOur open-source tracking software provides rich behavioral phenotyping of animals in many behavioral tasks. The flexible ROI design and live tracking makes the software adaptable to diverse paradigms.
Kang, X.; Yu, T.; Xu, K.; Liu, C.; Wu, R.
Show abstract
With the rapid development of Large Language Models (LLMs) and Agent technologies, AI can assist in solving a variety of real-world problems across multiple domains, such as autonomous driving, drug discovery, and materials design. In this work, we present EnzySeek, an enzyme catalysis AI agent designed to assist researchers in enzyme catalysis simulations. First, we constructed a domain-specific knowledge base by curating thousands of papers related to enzyme catalysis. Second, we customized Model Context Protocol (MCP) interfaces for each step of the enzyme catalysis simulation workflow, enabling these functions to be invoked by LLMs. Finally, we configured an agent capable of simultaneously referencing past empirical studies on enzyme catalysis, autonomously executing tool calls, and analyzing as well as presenting the results. EnzySeeks capabilities cover multiple aspects, including protein structure prediction, molecular docking, system preparation and parameterization, molecular dynamics (MD) simulations, and QM/MM calculations. The conclusions drawn by EnzySeek are primarily based on the results of QM/MM calculations. We employed the semi-empirical quantum mechanical method GFN2-xTB to calculate the QM region of the system. Benchmark results indicate that the GFN2-xTB method can achieve high efficiency while maintaining accuracy. The EnzySeek agent is designed to continuously learn from newly published literature and past computational tasks. During its operation, every AI decision is manually verified and scored by human experts. This human-in-the-loop validation provides the AI with sufficient case-based support, ultimately contributing to the full automation of enzyme catalysis computations. All data generated during the simulations are compiled into a dataset, which is used to establish evaluation criteria specific to enzyme catalysis computational results.
Zougman, A.
Show abstract
The protein sample preparation methods for shotgun proteomics are nowadays well-established unlike the ones for whole protein analysis. The goal of my work has been to create a simple methodology which provides a single uncomplicated sample preparation tool for these two fields. Nowadays the bulk of proteomics work is done using detergents for protein solubilization. The presented concept, which is based on unspecific adsorption of protein molecules on wide pore materials, allows for protein capture and clean-up from solutions of the most commonly used sodium dodecyl sulfate detergent. It could also be applied to proteins in detergent-free solutions. After the capture and clean-up, proteins could be either cleaved for the downstream peptide analysis or eluted for the whole protein analysis. If required, the eluted whole proteins could be recaptured and cleaved into peptides. Depending on the experimental goals, the sample preparation device could be fitted with embedded proteolytic enzymes to simplify routine sample processing and/or reversed phase media for the downstream peptide or protein separation.
Roberge, H.; Woller, T.; Pavie, B.; Hennies, J.; de Heus, C.; Edakkandiyil, L.; Liv, N.; Munck, S.
Show abstract
Correlative Light and Electron Microscopy (CLEM) integrates the molecular specificity of light microscopy (LM) with the ultrastructural detail of electron microscopy (EM), enabling comprehensive spatial analysis of biological samples. Despite growing demand, processing 3D CLEM datasets remains challenging, specifically for service provision in facilities, due to their multimodal nature and the lack of unified approaches. Typical steps include EM slice alignment, LM-EM registration, segmentation, and 3D visualization. We present a modular, end-to-end pipeline that consolidates existing and newly developed tools into a coherent workflow for 3D CLEM analysis and allows railroading the approach. Designed as interoperable modules accessible through a user-friendly interface, the pipeline is fully open-source and scales from standard workstations to high-performance computing environments to address the need for analysis of growing datasets. While some steps still require manual input, individual components can be automated to increase throughput and reproducibility. Together, this integrated solution lowers technical barriers and supports broader adoption of 3D CLEM methodologies.
Walker, a.; Guberman-Pfeffer, M. J.
Show abstract
Millions of experimental and AI-predicted protein structures are now available, and the biosynthetic promise of bespoke proteins is increasingly within reach. The functional characterization challenge thus posed cannot be addressed by experimental techniques alone. Molecular dynamics (MD) simulations offer functional screening with atomic resolution, yet accessibility remains limited. Existing computational chemistry software presents stark trade-offs whereby powerful tools require extensive expertise and manual effort, or user-friendly programs function as black boxes that obscure critical preparation decisions. Herein, we present ProPrep, an interactive workflow manager that guides users through expert-quality MD preparation by showing the what, why, and how of each step while automating tedious manual operations. Within a single workspace, ProPrep integrates (1) downloading structures from multiple sources (PDB, AlphaFold, AlphaFill), (2) performing homology searches, (3) aligning structures, (4) curating and repairing structural issues, (5) applying mutations, (6) parameterizing specialized residues, (7) converting redox-active sites to forcefield-compatible forms, (8) generating topology and coordinate files, and (9) configuring, executing, and analyzing simulations with active monitoring of key quantities via ASCII visualizations. A key innovation is ProPreps extensible transformer framework for detecting, defining, and transforming redox-active sites--including mono- and polynuclear metal centers, organic cofactors, and redox-active amino acids--for forcefield compatibility. We demonstrate the full workflow on a 64-heme cytochrome nanowire bundle (PDB: 9YUQ), proceeding from a PDF file to energy minimization of the solvated system (467,635 atoms) for constant pH molecular dynamics--a process demanding 4,819 PDB record modifications and 610 bond definitions--in 18 minutes of user interaction. The entire process is recorded in an interactive session log that can be shared and replayed for reproducibility, making simulation setup a fully transparent process that relies on what was done instead of what was remembered and reported.
Antony, F.; Bhattacharya, A.; Duong van Hoa, F.
Show abstract
Peptergent is a novel class of amphipathic peptides that enable detergent-free extraction and purification of membrane proteins (MPs). These designed peptides self-assemble around hydrophobic transmembrane regions of proteins, forming stable, water-soluble assemblies that can be isolated directly from biological membranes. By doing so, Peptergent bypass the limitations imposed by traditional detergents, which often destabilize proteins and restrict downstream analyses. Since detergents are completely avoided, Peptergent-isolated MPs are directly amenable to structural and mass spectrometry (MS) analysis, thereby addressing their persistent underrepresentation in proteomic datasets and improving their accessibility for drug-screening strategies. Here, we describe a streamlined protocol for isolating MPs with the Peptergent PDET-1, followed by exchange into His-tagged Peptidiscs for Ni-NTA-based affinity purification. The method comprises membrane isolation, peptide preparation, protein extraction, clarification, and exchange of MPs from Peptergent to Peptidiscs. Application of this workflow yields enriched membrane proteomes compatible with downstream LC-MS/MS analysis, with improved recovery of hydrophobic and multi-pass membrane proteins. Key featuresO_LIDirect extraction and solubilization of membrane proteins in Peptergents C_LIO_LIExchange into His-tagged Peptidiscs enabling affinity purification of MPs C_LIO_LI100% detergent-free workflow compatible with LC-MS/MS analysis C_LIO_LIApplicable to cultured cells and tissue-derived membrane fractions C_LI In BriefWe describe a Peptergent-based workflow for isolating membrane proteins directly from membrane preparations. Proteins are extracted with the Peptergent peptide scaffold (PDET-1) and transferred into His-tagged Peptidisc (HD-43). The water-soluble membrane proteins are enriched by Ni-NTA affinity purification and prepared for bottom-up mass spectrometry, yielding enriched membrane proteomes and dried peptide samples ready for LC-MS analysis Graphical Overview O_FIG O_LINKSMALLFIG WIDTH=200 HEIGHT=128 SRC="FIGDIR/small/711971v1_ufig1.gif" ALT="Figure 1"> View larger version (36K): org.highwire.dtl.DTLVardef@af3241org.highwire.dtl.DTLVardef@c6a94org.highwire.dtl.DTLVardef@129322aorg.highwire.dtl.DTLVardef@19c7c9d_HPS_FORMAT_FIGEXP M_FIG C_FIG