Back

AMPBAN: A Deep Learning Framework Integrating Protein Sequence and Structural Features for Antimicrobial Peptide Prediction

Bai, W.; Yang, W.; Chen, Y.-Q.; Ji, H.; Brennan, L.; Wang, L.

2026-01-22 bioinformatics
10.64898/2026.01.20.700468 bioRxiv
Show abstract

The escalating crisis of antimicrobial resistance poses a devastating and immediate threat to human life. Antimicrobial peptides (AMPs) are a promising antibiotic substitute to combat antimicrobial resistance. Compared with the traditional wet-lab screening approaches, computational models have largely improved the efficiency of predicting antimicrobial peptide. However, most computational models overlook or underutilize the evolution and structural information of peptides, which is crucial for understanding the peptide functions. Here, we proposed a sophisticated deep learning model to predict AMPs, Antimicrobial Peptide Bilinear Attention Network (AMPBAN), which incorporates peptide evolution features from ESM3 protein language model, structure features from ESMFold predicted with equivariant graph neural network (EGNN), and the joint information from sequence and structure learned via Bilinear Attention Network. AMPBAN consistently demonstrated superior accuracy and generalization compared to nine state-of-the-art AMP prediction models across multiple independent benchmarks. Furthermore, an ablation study confirms that our multimodal fusion strategy significantly refines the integration of sequence and structural signals, yielding superior predictive balance over single-modality models. This framework provides a robust tool for the accelerated discovery of novel AMPs and the advancement of next-generation antimicrobial drug development. The datasets, source code and models are available at https://github.com/baiwenhuim/ampban.

Matching journals

The top 4 journals account for 50% of the predicted probability mass.

1
Briefings in Bioinformatics
326 papers in training set
Top 0.2%
14.4%
2
Advanced Science
249 papers in training set
Top 0.7%
14.1%
3
Journal of Chemical Information and Modeling
207 papers in training set
Top 0.5%
12.1%
4
Nature Machine Intelligence
61 papers in training set
Top 0.2%
9.9%
50% of probability mass above
5
Genomics, Proteomics & Bioinformatics
171 papers in training set
Top 2%
3.5%
6
Bioinformatics
1061 papers in training set
Top 6%
3.5%
7
Nature Communications
4913 papers in training set
Top 41%
3.5%
8
Computational and Structural Biotechnology Journal
216 papers in training set
Top 4%
1.9%
9
Communications Chemistry
39 papers in training set
Top 0.3%
1.7%
10
PLOS Computational Biology
1633 papers in training set
Top 17%
1.7%
11
Science Bulletin
22 papers in training set
Top 0.4%
1.6%
12
Nucleic Acids Research
1128 papers in training set
Top 11%
1.6%
13
Acta Pharmaceutica Sinica B
11 papers in training set
Top 0.4%
1.6%
14
Communications Biology
886 papers in training set
Top 11%
1.5%
15
National Science Review
22 papers in training set
Top 1%
1.5%
16
Chemical Science
71 papers in training set
Top 1%
1.3%
17
iScience
1063 papers in training set
Top 22%
1.2%
18
eLife
5422 papers in training set
Top 52%
0.9%
19
Quantitative Biology
11 papers in training set
Top 0.6%
0.9%
20
Cell Reports Physical Science
18 papers in training set
Top 0.7%
0.8%
21
Cell Discovery
54 papers in training set
Top 5%
0.8%
22
Science China Life Sciences
26 papers in training set
Top 2%
0.7%
23
Patterns
70 papers in training set
Top 3%
0.7%
24
Scientific Reports
3102 papers in training set
Top 77%
0.7%
25
Journal of Cheminformatics
25 papers in training set
Top 0.6%
0.7%
26
Proceedings of the National Academy of Sciences
2130 papers in training set
Top 46%
0.7%
27
Biomolecules
95 papers in training set
Top 3%
0.6%
28
IEEE Transactions on Computational Biology and Bioinformatics
17 papers in training set
Top 0.8%
0.6%
29
Nature Biomedical Engineering
42 papers in training set
Top 3%
0.6%