Back

MHCBind: A Pan- and Allele-Specific Model for Predicting Class I MHC-Peptide Binding Affinity

Peddi, N.; Bijjula, D. R.; Gogte, S.; Kondaparthi, V.

2026-03-23 bioinformatics
10.64898/2026.03.20.713120 bioRxiv
Show abstract

Major Histocompatibility Complex (MHC) molecules are essential to the immune system because they bind and present peptide antigens to T cells, enabling immune recognition and response. The specificity of MHC-peptide interactions is crucial for understanding immune-related diseases, developing personalized immunotherapies, and designing effective vaccines. Current computational methods, while powerful, often rely on a single type of molecular information, usually sequence, and implicitly model the interaction between the two molecules. To address these limitations, we introduce MHC-Bind, a novel deep learning framework that captures a more comprehensive and biologically relevant view of the binding event. MHCBinds architecture employs a dual-view feature extraction strategy for both the MHC and the peptide. A Graph Attention Network (GAT) learns topological features from predicted residue contact maps, while a parallel 1D Convolutional Neural Network (CNN) captures multi-scale patterns from sequence embeddings. These four distinct feature sets are then integrated in a cross-fusion module that uses an attention mechanism to model interactions between the two molecules. Finally, a multi-layer perceptron (MLP) regression head maps the fused interaction signature to a precise binding affinity score. In rigorous comparative benchmarks against recent variants, such as NetMHCpan, MHCFlurry, and MHCnuggets, MHCBind demonstrates superior performance, achieving a significantly lower average prediction error (RMSE: 0.1485) and a higher correlation (PCC: 0.7231) in allele-specific contexts. For pan-allele tasks, it excels at correctly ranking peptides with a superior Spearmans Correlation (SCC: 0.7102), a crucial advantage for practical applications. The frameworks design is inherently flexible, excelling in both allele-specific and pan-allele prediction tasks.

Matching journals

The top 7 journals account for 50% of the predicted probability mass.

1
Nature Machine Intelligence
61 papers in training set
Top 0.1%
14.4%
2
Bioinformatics
1061 papers in training set
Top 3%
10.1%
3
Nature Communications
4913 papers in training set
Top 26%
6.8%
4
Advanced Science
249 papers in training set
Top 2%
6.8%
5
Nucleic Acids Research
1128 papers in training set
Top 3%
6.4%
6
Cell Systems
167 papers in training set
Top 2%
4.9%
7
Briefings in Bioinformatics
326 papers in training set
Top 1%
4.9%
50% of probability mass above
8
Frontiers in Immunology
586 papers in training set
Top 2%
4.3%
9
PLOS Computational Biology
1633 papers in training set
Top 9%
4.0%
10
Science Advances
1098 papers in training set
Top 9%
2.7%
11
Communications Biology
886 papers in training set
Top 4%
2.6%
12
Computational and Structural Biotechnology Journal
216 papers in training set
Top 3%
2.4%
13
iScience
1063 papers in training set
Top 13%
1.8%
14
Genome Medicine
154 papers in training set
Top 5%
1.7%
15
Bioinformatics Advances
184 papers in training set
Top 3%
1.5%
16
Cell Genomics
162 papers in training set
Top 4%
1.3%
17
Scientific Reports
3102 papers in training set
Top 69%
1.0%
18
Genomics, Proteomics & Bioinformatics
171 papers in training set
Top 5%
0.9%
19
Cell Reports Medicine
140 papers in training set
Top 6%
0.9%
20
Journal of Chemical Information and Modeling
207 papers in training set
Top 3%
0.9%
21
Nature Biotechnology
147 papers in training set
Top 7%
0.8%
22
ImmunoInformatics
11 papers in training set
Top 0.2%
0.8%
23
eLife
5422 papers in training set
Top 55%
0.8%
24
mAbs
28 papers in training set
Top 0.3%
0.8%
25
Nature Methods
336 papers in training set
Top 6%
0.7%
26
Structure
175 papers in training set
Top 3%
0.7%
27
Patterns
70 papers in training set
Top 3%
0.7%
28
Proceedings of the National Academy of Sciences
2130 papers in training set
Top 46%
0.7%
29
Cell Reports Methods
141 papers in training set
Top 6%
0.6%
30
PLOS ONE
4510 papers in training set
Top 71%
0.6%