2021
DOI: 10.48550/arxiv.2106.00908
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

TransMIL: Transformer based Correlated Multiple Instance Learning for Whole Slide Image Classification

Abstract: Multiple instance learning (MIL) is a powerful tool to solve the weakly supervised classification in whole slide image (WSI) based pathology diagnosis. However, the current MIL methods are usually based on independent and identical distribution hypothesis, thus neglect the correlation among different instances. To address this problem, we proposed a new framework, called correlated MIL, and provided a proof for convergence. Based on this framework, we devised a Transformer based MIL (TransMIL), which explored … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
18
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 9 publications
(18 citation statements)
references
References 28 publications
(45 reference statements)
0
18
0
Order By: Relevance
“…Most of the current MIL methods assume that the instances in each bag are independent and identically distributed, thereby neglecting the correlation among different instances. Shao et al [239] present TransMIL to explore both morphological and spatial information in weakly supervised WSI classification. Specifically, TransMIL aggregates morphological information with two transformer-based modules and a position encoding layer as shown in Fig.…”
Section: Tumor Classificationmentioning
confidence: 99%
“…Most of the current MIL methods assume that the instances in each bag are independent and identically distributed, thereby neglecting the correlation among different instances. Shao et al [239] present TransMIL to explore both morphological and spatial information in weakly supervised WSI classification. Specifically, TransMIL aggregates morphological information with two transformer-based modules and a position encoding layer as shown in Fig.…”
Section: Tumor Classificationmentioning
confidence: 99%
“…Another example is mi-Net [45], which pools predictions from single instances to derive a bag-level prediction. Other architectures adopted to MIL scenarios includes capsule networks [46], transformer [39] and graph neural networks [44]. Moreover, many works focus on the attentionbased pooling operators, like AbMILP introduced in [20] that weights the instances embeddings to obtain a bag embedding.…”
Section: Multiple Instance Learningmentioning
confidence: 99%
“…Table 3: Our ProtoMIL achieves state-of-the-art results on the Camelyon16 dataset in terms of AUC metric, surpassing even the transformer-based architecture. Notice that values for comparison marked with "*" and "**" are taken from [25] and [39], respectively.…”
Section: Positive Bagmentioning
confidence: 99%
See 2 more Smart Citations