2023
DOI: 10.1101/2023.07.21.23292757
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Scaling Self-Supervised Learning for Histopathology with Masked Image Modeling

Abstract: Computational pathology is revolutionizing the field of pathology by integrating advanced computer vision and machine learning technologies into diagnostic workflows. Recently, self-supervised learning (SSL) has emerged as a promising solution to learn representations from histology patches, leveraging large volumes of unannotated whole slide images (WSI). In particular, Masked Image Modeling (MIM) showed remarkable results and robustness over purely contrastive learning methods. In this work, we explore the a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
4
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
3

Relationship

1
7

Authors

Journals

citations
Cited by 12 publications
(6 citation statements)
references
References 63 publications
(116 reference statements)
0
4
0
Order By: Relevance
“…Each module can be updated independently, allowing the entire framework to continuously benefit from advancements in their respective domains. We illustrate this by comparing various pre-trained tile-encoders as shown in Supplementary Table 1, notably the recent ctranspath network, and demonstrate that the WSI Giga-SSL representation benefits from an improved tile-encoder; a research effort that already receives much attention [1517].…”
Section: Resultsmentioning
confidence: 99%
“…Each module can be updated independently, allowing the entire framework to continuously benefit from advancements in their respective domains. We illustrate this by comparing various pre-trained tile-encoders as shown in Supplementary Table 1, notably the recent ctranspath network, and demonstrate that the WSI Giga-SSL representation benefits from an improved tile-encoder; a research effort that already receives much attention [1517].…”
Section: Resultsmentioning
confidence: 99%
“…We focused on building a robust, reproducible and fair benchmarking framework to compare different machine learning models built on bulk RNA-seq data with respect to their performance on different downstream tasks. The key elements used in the pipeline we developed (repeated holdout cross-validation and fixed HP tuning budget) are not new per-se (Filiot et al 2023) but to our knowledge this framework has been used in a limited number of studies (Smith et al 2020) evaluating ML models on omics data in cancer-specific downstream tasks. Tuning directly on a cross-validation without a held out test set could lead to inflated performance and justifies the need for a nested cross-validation framework or the proposed repeated holdout test set.…”
Section: Discussionmentioning
confidence: 99%
“…To reach its full potential, future work will be crucial to establish the use of the VirtualMultiplexer in real-world settings. From a technical standpoint, the generated virtual multiplexed stainings can enable the development of foundational models for IHC, as they have been successfully developed for brightfield H&E images [54][55][56]. However, developing such models requires high volumes of data, which is potentially challenging to acquire for IHC.…”
Section: Discussionmentioning
confidence: 99%