Back

Physics-Guided Deep Neural Networks: Correcting Physical Distortions in Protein Phase Separation Prediction

Wang, M.; Lu, T.; Song, Y.-h.; Li, y.

2026-04-21 cell biology
10.64898/2026.04.18.719364 bioRxiv
Show abstract

BackgroundIn computational biology, embedding known physical laws into deep learning models to construct "Physics-Informed Neural Networks" (PINNs) is a mainstream paradigm for enhancing model interpretability and extrapolation capability. However, in complex multi-physics coupling problems, there is a risk of competitive imbalance between the physical term and the flexible artificial intelligence (AI) residual term, causing the model to degenerate into a "black-box" fit and lose the original purpose of being physics-driven. MethodsIn this study, targeting the problem of predicting protein liquid-liquid phase separation (LLPS) behavior in response to environmental factors (temperature, salt concentration), we identified physical distortions, gradient vanishing, and numerical instability in the initial physics-AI hybrid model. Three core correction strategies were proposed: (1) Weight Allocation Logic Reconstruction: Force the physical trunk weight to 1.0 at the output layer, suppressing the AI residual term to the perturbation level of 0.05~0.1, ensuring physics dominance; (2) Robust Physics Formula Construction: Abandon the unstable power function and introduce a combination of Softplus and logarithmic functions to stably simulate the nonlinear effects of charge shielding; (3) Gain Compensation Alignment: Apply gain compensation to the weak signal branch (temperature) to ensure its effective participation in optimization. ResultsThe optimized model maintained a fitting accuracy of R2{approx}0.62 on the test set, while physical consistency was significantly enhanced. The model successfully restored the monotonic increase in solubility with temperature characteristic of UCST-type phase diagrams and correctly captured the nonlinear charge shielding features in the salt concentration response. The weights of key physical parameters (e.g., hydrophobic contribution w_h, net charge contribution w_ncpr) increased from <10-3 to the 10-2 magnitude, demonstrating the reactivation of the physical branch. ConclusionsThe weight control, formula stabilization, and signal gain alignment strategies proposed in this study effectively address the classic problem of "AI hijacking" physics in physics-AI hybrid models. This work provides a universal solution for constructing biophysical predictive models that combine high fitting accuracy with strong physical interpretability.

Matching journals

The top 10 journals account for 50% of the predicted probability mass.

1
Computers in Biology and Medicine
120 papers in training set
Top 0.1%
15.3%
2
PLOS Computational Biology
1633 papers in training set
Top 5%
6.6%
3
PLOS ONE
4510 papers in training set
Top 26%
6.6%
4
Physical Biology
43 papers in training set
Top 0.3%
4.1%
5
Computational and Structural Biotechnology Journal
216 papers in training set
Top 1%
3.7%
6
BMC Bioinformatics
383 papers in training set
Top 3%
3.7%
7
Scientific Reports
3102 papers in training set
Top 33%
3.7%
8
Bioinformatics
1061 papers in training set
Top 6%
2.4%
9
Briefings in Bioinformatics
326 papers in training set
Top 3%
2.2%
10
Nature Machine Intelligence
61 papers in training set
Top 2%
1.8%
50% of probability mass above
11
Biology Methods and Protocols
53 papers in training set
Top 0.8%
1.8%
12
Frontiers in Plant Science
240 papers in training set
Top 3%
1.8%
13
International Journal of Biological Macromolecules
65 papers in training set
Top 2%
1.5%
14
Computational Biology and Chemistry
23 papers in training set
Top 0.2%
1.5%
15
Heliyon
146 papers in training set
Top 2%
1.5%
16
npj Systems Biology and Applications
99 papers in training set
Top 1%
1.4%
17
F1000Research
79 papers in training set
Top 2%
1.4%
18
iScience
1063 papers in training set
Top 19%
1.4%
19
Journal of Chemical Information and Modeling
207 papers in training set
Top 2%
1.1%
20
Frontiers in Cell and Developmental Biology
218 papers in training set
Top 7%
0.9%
21
Bioengineering
24 papers in training set
Top 0.9%
0.9%
22
PROTEOMICS
35 papers in training set
Top 0.6%
0.8%
23
Chemical Engineering Journal
10 papers in training set
Top 0.5%
0.8%
24
Patterns
70 papers in training set
Top 2%
0.8%
25
SoftwareX
15 papers in training set
Top 0.3%
0.8%
26
PeerJ
261 papers in training set
Top 13%
0.8%
27
ACS Omega
90 papers in training set
Top 3%
0.8%
28
Biophysical Journal
545 papers in training set
Top 5%
0.8%
29
Frontiers in Pharmacology
100 papers in training set
Top 4%
0.8%
30
eLife
5422 papers in training set
Top 56%
0.8%