Back

Sequence-aware Prediction of Point Mutation-induced Effects on Protein-Protein Binding Affinity using Deep Learning

Zhuang, J.; Li, Z.; Wang, S.; Zheng, R.; Zhang, G.

2025-11-16 bioinformatics
10.1101/2025.11.15.688659 bioRxiv
Show abstract

Amino acid mutations may lead to significant changes in the binding affinity of protein complexes, thereby causing a series of cellular dysfunctions. Therefore, accurate prediction of protein-protein binding affinity changes ({Delta}{Delta}G) induced by amino acid mutations is of great importance for understanding protein-protein interactions (PPIs). In this study, we propose SAMAffinity, a protein sequence-aware deep learning architecture for predicting changes in protein-protein binding affinity caused by amino acid mutations. SAMAffinity predicts mutation-induced {Delta}{Delta}G by integrating multi-source sequence features, leveraging a Mutation-Site Identification (MSI) module to highlight local semantic shifts and a Binding-Interface Awareness (BIA) module to capture interaction changes. Benchmark evaluations on public datasets show that under the mutation-level data splitting strategy, SAMAffinity outperforms the state-of-the-art sequence-based method AttABseq by 33.3%, 72.3%, 31.8%, and 30.5% on S1131, S4169, S645, and M1101 datasets, respectively. Moreover, under the complex-level data splitting strategy, SAMAffinity surpasses the structure-based method MpbPPI by 22.9%, 22.7%, 5.0%, and 11.4% on the corresponding datasets. Beyond predictive accuracy, the strong consistency between the models predicted distribution and natural amino-acid mutation tendencies indicates that SAMAffinity effectively captures the underlying mutational landscape shaped by intrinsic biochemical and evolutionary factors. Based on this capability, SAMAffinity demonstrated strong generalization in a study of severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) cases, suggesting its potential for optimizing therapeutic antibody design.

Matching journals

The top 5 journals account for 50% of the predicted probability mass.

1
Briefings in Bioinformatics
326 papers in training set
Top 0.1%
27.8%
2
Bioinformatics
1061 papers in training set
Top 3%
8.4%
3
Nature Machine Intelligence
61 papers in training set
Top 0.4%
6.3%
4
Computational and Structural Biotechnology Journal
216 papers in training set
Top 0.8%
4.9%
5
Genomics, Proteomics & Bioinformatics
171 papers in training set
Top 1%
4.9%
50% of probability mass above
6
Nature Communications
4913 papers in training set
Top 35%
4.3%
7
Advanced Science
249 papers in training set
Top 5%
4.0%
8
Journal of Chemical Information and Modeling
207 papers in training set
Top 1%
3.6%
9
PLOS Computational Biology
1633 papers in training set
Top 13%
2.5%
10
IEEE Journal of Biomedical and Health Informatics
34 papers in training set
Top 0.9%
1.9%
11
Computers in Biology and Medicine
120 papers in training set
Top 2%
1.8%
12
Communications Biology
886 papers in training set
Top 8%
1.7%
13
National Science Review
22 papers in training set
Top 0.9%
1.7%
14
Cell Systems
167 papers in training set
Top 8%
1.3%
15
IEEE Transactions on Computational Biology and Bioinformatics
17 papers in training set
Top 0.3%
1.3%
16
Frontiers in Immunology
586 papers in training set
Top 5%
1.2%
17
Nucleic Acids Research
1128 papers in training set
Top 14%
1.2%
18
NAR Genomics and Bioinformatics
214 papers in training set
Top 3%
1.2%
19
Scientific Reports
3102 papers in training set
Top 66%
1.2%
20
Bioinformatics Advances
184 papers in training set
Top 4%
1.0%
21
PLOS ONE
4510 papers in training set
Top 64%
0.9%
22
Science Bulletin
22 papers in training set
Top 0.7%
0.8%
23
iScience
1063 papers in training set
Top 29%
0.8%
24
Journal of Molecular Biology
217 papers in training set
Top 3%
0.8%
25
Journal of Genetics and Genomics
36 papers in training set
Top 2%
0.7%
26
eLife
5422 papers in training set
Top 59%
0.7%
27
Patterns
70 papers in training set
Top 3%
0.7%
28
Expert Systems with Applications
11 papers in training set
Top 0.6%
0.6%
29
Quantitative Biology
11 papers in training set
Top 0.9%
0.6%