2021
DOI: 10.1007/978-3-030-86520-7_27
|View full text |Cite
|
Sign up to set email alerts
|

Probing Negative Sampling for Contrastive Learning to Learn Graph Representations

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
34
1

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 14 publications
(35 citation statements)
references
References 9 publications
0
34
1
Order By: Relevance
“…However, the generalization bounds there enjoy a linear dependency on k, which would not be effective if k is large. Moreover, this is not consistent with many studies which show a large number of negative examples [11,21,23,38] is necessary for good generalization performance. For example, the work [20] used 65536 negative examples in unsupervised visual representation learning, for which the existing analysis requires n ≥ (65536) 2 d training examples to get non-vacuous bounds [3].…”
Section: Introductioncontrasting
confidence: 72%
See 2 more Smart Citations
“…However, the generalization bounds there enjoy a linear dependency on k, which would not be effective if k is large. Moreover, this is not consistent with many studies which show a large number of negative examples [11,21,23,38] is necessary for good generalization performance. For example, the work [20] used 65536 negative examples in unsupervised visual representation learning, for which the existing analysis requires n ≥ (65536) 2 d training examples to get non-vacuous bounds [3].…”
Section: Introductioncontrasting
confidence: 72%
“…The performance of machine learning (ML) models often depends largely on the representation of data, which motivates a resurgence of contrastive representation learning (CRL) to learn a representation function f : X → R d from unsupervised data [11,20,23]. The basic idea is to pull together similar pairs (x, x + ) and push apart disimilar pairs (x, x − ) in an embedding space, which can be formulated as minimizing the following objective [11,32]…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Our CL and SCL losses enforce consistency of the network across subsequent training epochs, which favors invariance of the network outputs to the data augmentation. This can be connected to recent trends of self-supervised learning for instance discrimination, ensuring that the embeddings of data-augmented versions of an instance are closer in embedding space than the embeddings of different instances [48,36,18,20,6]. In the fully annotated multilabel image classification setting, [15] encourages consistency of the spatial activations of the network among two data augmentations of an image, akin to a spatial extension of the Π-model [29].…”
Section: Related Workmentioning
confidence: 99%
“…Self-supervised learning (SSL) is a learning paradigm that presents itself as a more scalable approach to pretrain models for transfer learning with less human labeled data [2,3,9,11]. SSL can be understood as a two-step approach: a) learning data representation from solving a proxy or pretext task using automatically generated pseudolabels from raw-unlabeled data, followed by b) fine-tuning of the learned features on the actual downstream task, i.e.…”
Section: Related Workmentioning
confidence: 99%