2024
DOI: 10.1038/s41591-024-02857-3
|View full text |Cite
|
Sign up to set email alerts
|

Towards a general-purpose foundation model for computational pathology

Richard J. Chen,
Tong Ding,
Ming Y. Lu
et al.
Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
10
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 41 publications
(10 citation statements)
references
References 145 publications
0
10
0
Order By: Relevance
“…The HistoTME model consists of two main components: a frozen feature extraction component and a trainable attention-based multiple instance learning (AB-MIL) component 33 . In our efforts to efficiently train HistoTME, we explored three state-of-the-art open-source foundational models— CTransPath, RetCCL, and UNI 25,26,30 —as potential feature extractors. Additionally, we conducted experiments with two distinct approaches for AB-MIL: a single-task approach featuring a unique attention and multilayer perceptron (MLP) head for each TME signature, and a multitask approach, which incorporates a shared attention head for functionally related TME signatures but maintains separate MLP heads for each individual TME signature (Fig.…”
Section: Resultsmentioning
confidence: 99%
See 2 more Smart Citations
“…The HistoTME model consists of two main components: a frozen feature extraction component and a trainable attention-based multiple instance learning (AB-MIL) component 33 . In our efforts to efficiently train HistoTME, we explored three state-of-the-art open-source foundational models— CTransPath, RetCCL, and UNI 25,26,30 —as potential feature extractors. Additionally, we conducted experiments with two distinct approaches for AB-MIL: a single-task approach featuring a unique attention and multilayer perceptron (MLP) head for each TME signature, and a multitask approach, which incorporates a shared attention head for functionally related TME signatures but maintains separate MLP heads for each individual TME signature (Fig.…”
Section: Resultsmentioning
confidence: 99%
“…Due to the limited benchmarking of foundation models for continuous biomarker prediction tasks, we experimented with three popular foundation models as feature extractors. We found out that the UNI foundation model 25 , when paired with multi-task AB-MIL, achieved the best predictive performance for the various TME signature prediction tasks. This improved performance likely stems from the considerably large histopathology datasets used to pre-train the UNI foundation model, as well as the added regularization induced by multitask learning.…”
Section: Discussionmentioning
confidence: 98%
See 1 more Smart Citation
“…Second, it remains challenging to design a model architecture that can effectively capture both local patterns in individual tiles and global patterns across whole slides 35 – 39 . Existing models often treat each image tile as an independent sample and formulate slide-level modelling as multiple instance learning 4 , 40 – 43 , thus limiting their ability to model complex global patterns in gigapixel whole slides. A notable exception is Hierarchical Image Pyramid Transformer (HIPT), which explores hierarchical self-attention over the tiles 35 .…”
Section: Mainmentioning
confidence: 99%
“…Recently developed vision models allow unsupervised extraction of morphological features which can then be used for clustering and data integration tasks. During the hackathon general purpose models trained on imagenet and UNI (Chen et al, 2024) a model specifically tuned on histopathology were evaluated.…”
Section: Workgroup Spatial Multi-omicsmentioning
confidence: 99%