2022
DOI: 10.1101/2022.10.24.513504
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

The geometry of hidden representations of protein language models

Abstract: Protein language models (pLMs) transform their input into a sequence of hidden representations whose geometric behavior changes across layers. Looking at fundamental geometric properties such as the intrinsic dimension and the neighbor composition of these representations, we observe that these changes highlight a pattern characterized by three distinct phases. This phenomenon emerges across many models trained on diverse datasets, thus revealing a general computational strategy learned by pLMs to reconstruct … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
6
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
3
1

Relationship

1
6

Authors

Journals

citations
Cited by 10 publications
(15 citation statements)
references
References 25 publications
(45 reference statements)
0
6
0
Order By: Relevance
“…More generally, the last hidden layer has been optimized for the un-supervised pre-training objective (learning to reproduce masked sequences). This optimization might not be best for any downstream task 56,57 . Fine-tuning may exactly address this issue by optimizing the models output for particular tasks.…”
Section: Discussionmentioning
confidence: 99%
“…More generally, the last hidden layer has been optimized for the un-supervised pre-training objective (learning to reproduce masked sequences). This optimization might not be best for any downstream task 56,57 . Fine-tuning may exactly address this issue by optimizing the models output for particular tasks.…”
Section: Discussionmentioning
confidence: 99%
“…By analyzing the intrinsic dimension (ID) in transformers, Valeriani et al (2023) discovered that data initially spans a high-dimensional manifold in the earliest layers but then undergoes considerable contraction. These insights pave the way for potentially differentiated strategies in handling manifold representations at various depths of a network.…”
Section: Discussionmentioning
confidence: 99%
“…The aim of this study is to provide researchers with a database that could accelerate the identification of novel protein families in the Human Gastrointestinal Proteome, thus potentially enabling the formulation of new functional hypotheses. Future plans include further optimization of the clustering procedure allowing for faster processing of newly deposited metagenomic data, and combining automatic classification with functional annotation leveraging deep learning 21,22 for even more comprehensive annotation of UHGP.…”
Section: Background and Summarymentioning
confidence: 99%