The platform will undergo maintenance on Sep 14 at about 7:45 AM EST and will be unavailable for approximately 2 hours.
2022
DOI: 10.1101/2022.11.14.516459
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

MOT: a Multi-Omics Transformer for multiclass classification tumour types predictions

Abstract: Motivation Breakthroughs in high-throughput technologies and machine learning methods have enabled the shift towards multi-omics modelling as the preferred means to understand the mechanisms underlying biological processes. Machine learning enables and improves complex disease prognosis in clinical settings. However, most multi-omic studies primarily use transcriptomics and epigenomics due to their over-representation in databases and their early technical maturity compared to others omics. For complex phenoty… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 42 publications
0
3
0
Order By: Relevance
“…However, incorporating multi-omics data and their crosstalk information in a Transformer is very challenging: when processing multi-omics data, the multi-modal features are usually multiplied by tens of thousands of genes, producing an extremely long input that is not acceptable by a common Transformer model (usually <512 words). Meanwhile, certain embedding methods for biological data, such as discretization and linear transformation, were introduced in the previous Transformer models ( Osseni et al 2022 , Cui et al 2023 , Theodoris et al 2023 ), while biological information was largely lost during these kinds of embedding.…”
Section: Introductionmentioning
confidence: 99%
“…However, incorporating multi-omics data and their crosstalk information in a Transformer is very challenging: when processing multi-omics data, the multi-modal features are usually multiplied by tens of thousands of genes, producing an extremely long input that is not acceptable by a common Transformer model (usually <512 words). Meanwhile, certain embedding methods for biological data, such as discretization and linear transformation, were introduced in the previous Transformer models ( Osseni et al 2022 , Cui et al 2023 , Theodoris et al 2023 ), while biological information was largely lost during these kinds of embedding.…”
Section: Introductionmentioning
confidence: 99%
“…However, incorporating multi-omics data and their crosstalk information in a Transformer is very challenging: when processing multi-omics data, the multi-modal features are usually multiplied by the number of genes (tens of thousands), producing an extremely long input that is not acceptable by a common Transformer model (usually less than 512). Meanwhile, certain embedding methods for biological data, such as discretization and linear transformation, were introduced in the previous Transformer models [17][18][19] , while biological information was largely lost during these kinds of embedding.…”
Section: Introductionmentioning
confidence: 99%
“…The crisscross attention mechanism of Transformer would be very useful to capture the crosstalk information 15 . However, the current embedding methods for biological data in Transformer models [16][17][18] , such as discretization and linear transformation, usually missed the biological information. Meanwhile, multi-modal features were typically multiplied by the number of genes, resulting in long input sequences, which could strain available memory resources or necessitate a feature selection step 18,19 .…”
Section: Introductionmentioning
confidence: 99%