3260 papers • 126 benchmarks • 313 datasets
Image credit: FastFold: Reducing AlphaFold Training Time from 11 Days to 67 Hours
(Image credit: Papersgraph)
These leaderboards are used to track progress in protein-structure-prediction-12
Use these libraries to find protein-structure-prediction-12 models and implementations
This work validated an entirely redesigned version of the neural network-based model, AlphaFold, in the challenging 14th Critical Assessment of protein Structure Prediction (CASP14)15, demonstrating accuracy competitive with experimental structures in a majority of cases and greatly outperforming other methods.
This work proposes a deep convolutional neural network architecture, MUST-CNN, that uses a novel multilayer shift-and-stitch technique to generate fully dense per-position predictions on protein sequences and beats the state-of-the-art performance on two large protein property prediction datasets.
This work demonstrates state-of-the-art protein structure prediction (PSP) results using embeddings and deep learning models for prediction of backbone atom distance matrices and torsion angles, and creates a new gold standard dataset of proteins which is comprehensive and easy to use.
This work shows how to generate set-valued predictions from a black-box predictor that controls the expected loss on future test points at a user-specified level, and provides explicit finite-sample guarantees for any dataset by using a holdout set to calibrate the size of the prediction sets.
The ProteinNet series of data sets were created to provide a standardized mechanism for training and assessing data-driven models of protein sequence-structure relationships and to create validation sets distinct from the official CASP sets that faithfully mimic their difficulty.
An iterative version of the SE(3)-Transformer is implemented, an SE( 3)-equivariant attention-based model for graph data, to address the additional complications which arise when applying the SE (3)- Transformer in an iterative fashion and consider why a iterative model may be beneficial in some problem settings.
ProteinBERT is introduced, a deep language model specifically designed for proteins that obtains state-of-the-art performance on multiple benchmarks covering diverse protein properties, despite using a far smaller model than competing deep-learning methods.
This work presents the first million-level protein structure prediction dataset with high coverage and diversity, named as PSP, and provides in addition the benchmark training procedure for SOTA protein structure prediction model on this dataset.
Adding a benchmark result helps the community track progress.