ESMDisPred: A Structure-Aware CNN-Transformer Architecture for Intrinsically Disordered Protein Prediction

This article has been Reviewed by the following groups

Read the full article See related articles

Discuss this preprint

Start a discussion What are Sciety discussions?

Listed in

Log in to save this article

Abstract

Intrinsically disordered proteins (IDPs) lack stable three-dimensional structures, yet play vital roles in key biological processes, including signaling, transcription regulation, and molecular scaffolding. Their structural flexibility presents significant challenges for experimental characterization and contributes to diseases such as cancer and neurodegenerative disorders. Accurate computational prediction of IDPs is important for advancing research and drug discovery, structural biology, and protein engineering. In this study, we introduce ESMDisPred, a novel structure-aware disorder predictor that builds on the representational power of Evolutionary Scale Modeling-2 (ESM2) protein language models. ESMDisPred integrates sequence embeddings with structural information from the Protein Data Bank (PDB) to deliver state-of-the-art prediction accuracy. Model performance is further enhanced through feature engineering strategies, including terminal residue encoding, statistical summarization, and sliding-window analysis. To capture both local sequence motifs and long-range dependencies, we designed a hybrid CNN-Transformer architecture that balances convolutional efficiency with the representational power of self-attention. On CAID3 benchmarks, our latest model achieves ROC-AUC 0.895, AP 0.778, and a max F1 of 0.759, outperforming recent methods. Our results highlight the importance of integrating protein language model embeddings with explicit structural information for improved disorder prediction.

Article activity feed

  1. To reduce class imbalance, we exclude these longer sequences from the training dataset.

    Downsampling/excluding a minority class is an interesting decision. Why not include some of them, use some flavor of stratified sampling/curriculum learning/train a specialized subnetwork/set of heads on the larger sequences with a reasonable split? How does the model generalize/perform on larger sequences?

  2. In our earlier work, we introduced DisPredict3.0, the most recent iteration of the DisPredict series, which integrates evolutionary representations derived from protein language models to improve the prediction of intrinsically disordered regions (IDRs) [5]. This approach achieved the top ranking on the Disorder NOX dataset in CAID2. Building on this foundation, we now present ESMDisPred, a structure-aware disordered protein predictor that incorporates embeddings from the Evolutionary Scale Modeling-2 (ESM2) language model [3]. ESM2 is considered the SOTA language model and has demonstrated exemplary performance in protein structure prediction (ESMFold)

    This is interesting. Evolutionary context can be really informative?