Proceedings of the 11th ACM International Conference on Bioinformatics, Computational Biology and Health Informatics 2020
DOI: 10.1145/3388440.3412467
|View full text |Cite
|
Sign up to set email alerts
|

Transforming the Language of Life

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
23
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 75 publications
(25 citation statements)
references
References 31 publications
1
23
0
Order By: Relevance
“…Another viable option are the recurrent and attention-based neural networks, which have enough computational power to describe relevant dependencies in protein sequences [108, 109, 110]. However, while modern neural networks have been successfully applied to annotation of protein families [111, 112], their performance in modeling short protein sequence fragments is yet too be evaluated.…”
Section: Discussionmentioning
confidence: 99%
“…Another viable option are the recurrent and attention-based neural networks, which have enough computational power to describe relevant dependencies in protein sequences [108, 109, 110]. However, while modern neural networks have been successfully applied to annotation of protein families [111, 112], their performance in modeling short protein sequence fragments is yet too be evaluated.…”
Section: Discussionmentioning
confidence: 99%
“…Work in [8] debuts the PRoBERTa model. The model is pre-trained to learn task-agnostic sequence representations of amino-acid sequences.…”
Section: Related Workmentioning
confidence: 99%
“…In particular, they provide an alternative framework to link protein sequence to function without relying on sequence similarity. Sequence representations learned via PLMs have been shown useful for various prediction tasks, from predicting secondary structure [4], subcellular localization [4], [5], evolutionary relationships within protein families [6], superfamily [7], and family [8] membership.…”
Section: Introductionmentioning
confidence: 99%
“…For embedding protein sequences, Nambiar et al . [ 79 ] designed PRoBERTa, a neural network architecture based on RoBERTa. After pretraining on SWISS-PROT database and fine-tuning, their method performs well in protein family classification and protein interaction prediction.…”
Section: Application Of Pretraining Modelmentioning
confidence: 99%