2022
DOI: 10.21203/rs.3.rs-1379598/v1
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

OSR2Vec: an Ontology Semantic Representation Method Based Pre-training Natural Language Model

Abstract: Background: Ontology, as a formal description and organization for domain-specific knowledge, is served as a significant approach of information retrieval and biological discovery in biomedical research, attracting more and more attention in computational biomedical and bioinformatics. It is quantified entities using semantic similarity based on ontology terms when using ontologies for description or annotation. More recently, machine learning approaches have enabled ontologies to describe concepts in ”computa… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
2
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 28 publications
0
2
0
Order By: Relevance
“…We select a variety of GNN pre-training baselines, which include the latest state-of-the-art methods for molecular pre-training: EdgePred ( Hamilton et al 2017 ), InfoMax ( Veličković et al 2019 ), ContextPred ( Hu et al 2020 ), AttrMask ( Hu et al 2020 ), GraphCL ( You et al 2020 ), GraphMAE ( Hou et al 2022 ), redGraphMVP ( Liu et al 2022 ), JOAO ( You et al 2021 ), JOAOv2 ( You et al 2021 ), SimGRACE ( Xia et al 2022 ), and Mole-BERT ( Xia et al 2023a ). In addition, we also consider a new graph contrast learning module, TMCL, proposed in Mole-BERT.…”
Section: Methodsmentioning
confidence: 99%
“…We select a variety of GNN pre-training baselines, which include the latest state-of-the-art methods for molecular pre-training: EdgePred ( Hamilton et al 2017 ), InfoMax ( Veličković et al 2019 ), ContextPred ( Hu et al 2020 ), AttrMask ( Hu et al 2020 ), GraphCL ( You et al 2020 ), GraphMAE ( Hou et al 2022 ), redGraphMVP ( Liu et al 2022 ), JOAO ( You et al 2021 ), JOAOv2 ( You et al 2021 ), SimGRACE ( Xia et al 2022 ), and Mole-BERT ( Xia et al 2023a ). In addition, we also consider a new graph contrast learning module, TMCL, proposed in Mole-BERT.…”
Section: Methodsmentioning
confidence: 99%
“…Nevertheless, the limited availability of labeled molecules and the vastness of the chemical space have constrained their prediction performance, particularly when handling out-of-distribution data samples 6,28,29 . Along with the remarkable achievements of selfsupervised learning methods in the fields of natural language processing 30,31 and computer vision 32,33 , these techniques have been employed to pre-train GNNs and improve the representation learning of molecules, leading to substantial improvements in the downstream molecular property prediction tasks 28,[34][35][36][37][38][39][40][41][42] .…”
mentioning
confidence: 99%
“…1a, the 3D shape is a promising molecular similarity metrics that contains more pharmacological and physical information than the 2D structure and fingerprint similarity, therefore, 3D shape and pharmacological similarity is an optimal metric for contrastive learning. In GraphMVP 5 , GEM 6 , and Uni-Mol 7 , the authors primarily focused on 3D conformation as the pre-training task. In contrast, MolCLaSS 16 incorporates static 3D shape similarity into the framework of contrastive learning.…”
mentioning
confidence: 99%