2022
DOI: 10.1101/2022.07.07.499217
|Get access via publisher |Cite
Preprint
|
Sign up to set email alerts

SNVformer: An Attention-based Deep Neural Network for GWAS Data

Abstract: Despite being the widely-used gold standard for linking common genetic variations to phenotypes and disease, genome-wide association studies (GWAS) suffer major limitations, partially attributable to the reliance on simple, typically linear, models of genetic effects. More elaborate methods, such as epistasis-aware models, typically struggle with the scale of GWAS data. In this paper, we build on recent advances in neural networks employing Transformer-based architectures to enable such models at a large scale… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2023
2023
2025
2025

Publication Types

Select...
3
1

Relationship

1
3

Authors

Journals

citations
Cited by 4 publications
(3 citation statements)
references
References 30 publications
0
3
0
Order By: Relevance
“…We use a variation of the SNVformer architecture [11], with additions to the SNV encoding. We tested several variations of the original SNVformer encoding (see Appendix F.2) the best-performing variation is described here.…”
Section: Methodsmentioning
confidence: 99%
“…We use a variation of the SNVformer architecture [11], with additions to the SNV encoding. We tested several variations of the original SNVformer encoding (see Appendix F.2) the best-performing variation is described here.…”
Section: Methodsmentioning
confidence: 99%
“…The complexities of genetic information pose unique challenges, such as high dimensionality and the need for significant computational power, which have so far hindered the widespread adoption of foundation models in this area with relatively few publications applying basic concepts of foundation models to genomic data [104,[106][107][108]. For example, Santiesteban et al [109] showed that foundation models combining transcriptomics and histopathology data through self-supervised learning significantly improve survival prediction.…”
Section: Opportunities Of Large Language Models and Foundation Modelsmentioning
confidence: 99%
“…Recent developments in deep learning [13][14][15] offer significant opportunities to advance the PRS framework [16][17][18][19] , as they have the capacity to model both complex epistatic interactions and knowledge of molecular mechanisms. Among these, the Transformer has shown strong potential to address longstanding problems in the biomedical sciences, including prediction of 3D protein structures [20][21][22] , biomedical image analysis 23 , inference of gene expression from genome sequence [24][25][26] , and mapping a sequence of SNPs to a predicted phenotype 27,28 . The Transformer architecture is known for its central use of an "attention" mechanism 29 , an operation that dynamically computes the importance of each input element relative to others, enabling the model to focus on the most relevant features [29][30][31] .…”
Section: Introductionmentioning
confidence: 99%