Enhancing RNA 3D Folding Prediction via Transformer and Equivariant Architectures under Resource Constraints

Read the full article See related articles

Listed in

This article is not in any list yet, why not save it to one of your lists.
Log in to save this article

Abstract

Predicting the three-dimensional structure of RNA molecules is a fundamental problem in computational biology, with applications in understanding gene regulation, viral function, and drug design. Existing RNA structure prediction methods often rely on large models or heavy computation, as exemplified by protein folding solutions like AlphaFold2 [2], which makes them impractical under tight resource constraints. In this work, we present an efficient neural architecture that integrates a Transformer encoder enhanced with Rotary Positional Embeddings (RoPE) and a small residual MLP block, along with an optional E(n)-equivariant refinement module. We train our model on the Stanford Ribonanza RNA 3D dataset [11] using BF16 precision on a single NVIDIA H100 (80 GB) GPU. Our model achieves lower RMSD than simpler baselines, demonstrating that modern Transformer and equivariant techniques can be effectively applied to RNA folding under practical compute constraints.

Article activity feed