2018
DOI: 10.1101/286096
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Gene2Vec: Distributed Representation of Genes Based on Co-Expression

Abstract: 1Background: Existing functional description of genes are categorical, discrete, and mostly

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
45
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 29 publications
(45 citation statements)
references
References 8 publications
0
45
0
Order By: Relevance
“…Both developmental brain gene expression profiles and RNA transcript sequence compositions were used as features for model construction. To reduce the high dimensionality of the input features, which might cause model overfitting, an autoencoder network, gene embedding [ 26 ], and RF-based feature selection were tested. Lastly, we utilized the models to predict and prioritize ASD-associated candidate lncRNAs, which might provide a good list of targets for further investigation.…”
Section: Introductionmentioning
confidence: 99%
“…Both developmental brain gene expression profiles and RNA transcript sequence compositions were used as features for model construction. To reduce the high dimensionality of the input features, which might cause model overfitting, an autoencoder network, gene embedding [ 26 ], and RF-based feature selection were tested. Lastly, we utilized the models to predict and prioritize ASD-associated candidate lncRNAs, which might provide a good list of targets for further investigation.…”
Section: Introductionmentioning
confidence: 99%
“…Specifically, bioVec [1] and seq2vec [5] have applied the word2vec technique to biological sequences. Similarly, Gene2vec [3], [21], and Dna2vec [13] applied the same technique to gene embedding, protein embedding, and DNA sequence embedding respectively. All these works were based on word2vec, and their k-mer sizes span from 3-8 (dna2vec embed k-mers of length 3 to 8, others work on k-mer size of 3 ).…”
Section: Related Workmentioning
confidence: 99%
“…This technique has been applied to a variety of subjects in biomedical sciences with interesting results. [16][17][18][19][20][21][22] A hallmark feature of word2vec is that the resulting embeddings encode semantic relationships. A classic example would be that, in embeddings trained on an English language corpus, the vectors going from countries to their capital are similar.…”
Section: Representation Of Medication Order Sequences As Word2vec Embmentioning
confidence: 99%