2024
DOI: 10.1038/s41467-024-51844-2
|View full text |Cite
|
Sign up to set email alerts
|

Fine-tuning protein language models boosts predictions across diverse tasks

Robert Schmirler,
Michael Heinzinger,
Burkhard Rost

Abstract: Prediction methods inputting embeddings from protein language models have reached or even surpassed state-of-the-art performance on many protein prediction tasks. In natural language processing fine-tuning large language models has become the de facto standard. In contrast, most protein language model-based protein predictions do not back-propagate to the language model. Here, we compare the fine-tuning of three state-of-the-art models (ESM2, ProtT5, Ankh) on eight different tasks. Two results stand out. First… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
0
0

Publication Types

Select...

Relationship

0
0

Authors

Journals

citations
Cited by 0 publications
references
References 54 publications
0
0
0
Order By: Relevance

No citations

Set email alert for when this publication receives citations?