2022
DOI: 10.48550/arxiv.2206.02353
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Beyond Just Vision: A Review on Self-Supervised Representation Learning on Multimodal and Temporal Data

Abstract: Recently, Self-Supervised Representation Learning (SSRL) has attracted much attention in the field of computer vision, speech, natural language processing (NLP), and recently, with other types of modalities, including time series from sensors. The popularity of self-supervised learning is driven by the fact that traditional models typically require a huge amount of well-annotated data for training. Acquiring annotated data can be a difficult and costly process. Self-supervised methods have been introduced to i… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2
1

Relationship

1
5

Authors

Journals

citations
Cited by 6 publications
(6 citation statements)
references
References 80 publications
0
6
0
Order By: Relevance
“…In the field of healthcare and physiological applications, SSL faces additional challenges due to the heterogeneous nature of data acquired from different sensors with varying characteristics, sampling rates, and resolutions (Deldari et al, 2022a). The dynamic nature of real-world situations further complicates the aggregation and compression of multimodal sensor data into a coherent global embedding suitable for downstream tasks.…”
Section: Icml Workhop On Machinementioning
confidence: 99%
“…In the field of healthcare and physiological applications, SSL faces additional challenges due to the heterogeneous nature of data acquired from different sensors with varying characteristics, sampling rates, and resolutions (Deldari et al, 2022a). The dynamic nature of real-world situations further complicates the aggregation and compression of multimodal sensor data into a coherent global embedding suitable for downstream tasks.…”
Section: Icml Workhop On Machinementioning
confidence: 99%
“…This section focuses on approaches that are compatible with biosignals and that were encountered during the survey. Various classification schemes have been proposed for pretext tasks, depending on the domain of application [33]. Here, methodologies will be grouped in the following categories: predictive pretext tasks, generative pretext tasks, contrastive learning pretext tasks.…”
Section: A Pretext Tasksmentioning
confidence: 99%
“…We have demonstrated the feasibility of predicting impaired MFR using a relatively simple CNN with only 72,929 trainable parameters. We expect that CNN model performance may be further improved by additional machine learning technologies such as larger CNN architectures (41), self-supervised model pre-training (62), and transformer-based frameworks (40,63). A further possible extension would be to train the CNN using only rest ECG waveforms or reduced leads which could enable use of ambulatory ECG monitoring data.…”
Section: Future Studiesmentioning
confidence: 99%