2023
DOI: 10.1109/access.2023.3245023
|View full text |Cite
|
Sign up to set email alerts
|

Modality Specific CBAM-VGGNet Model for the Classification of Breast Histopathology Images via Transfer Learning

Abstract: Histopathology images are very distinctive, and one image may contain thousands of objects.Transferring features from natural images to histopathology images may not provide impressive outcomes. In this study, we have proposed a novel modality-specific CBAM-VGGNet model for classifying H and E stained breast histopathology images. Instead of using pre-trained models on ImageNet, we have trained VGG16 and VGG19 models on the same domain cancerous histopathology datasets which are then used as fixed feature extr… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
8
1

Relationship

0
9

Authors

Journals

citations
Cited by 17 publications
(5 citation statements)
references
References 55 publications
(44 reference statements)
0
5
0
Order By: Relevance
“…Some literature, such as the work by Joseph et al 16 , reports higher performance metrics; however, these results may be subject to overfitting, casting doubt on their generalizability. Similarly, other studies, like that of Ijaz et al 25 , have applied their models at only one magnification level, which may limit the applicability of their findings across varying conditions. These observations underscore the importance of cautious interpretation of comparative performance metrics and highlight the need for comprehensive testing across diverse conditions to ensure robust and reliable model performance.…”
Section: Discussionmentioning
confidence: 98%
See 1 more Smart Citation
“…Some literature, such as the work by Joseph et al 16 , reports higher performance metrics; however, these results may be subject to overfitting, casting doubt on their generalizability. Similarly, other studies, like that of Ijaz et al 25 , have applied their models at only one magnification level, which may limit the applicability of their findings across varying conditions. These observations underscore the importance of cautious interpretation of comparative performance metrics and highlight the need for comprehensive testing across diverse conditions to ensure robust and reliable model performance.…”
Section: Discussionmentioning
confidence: 98%
“…Ijaz et al 25 introduced the CBAM-VGGNet model, a fusion of VGG16 and VGG19, specifically trained on cancerous histopathology datasets. The model's complexity was streamlined using the GAP layer and CBAM, ensuring a focus on vital features.…”
Section: Related Workmentioning
confidence: 99%
“…Remarkably, the model pre-trained on the breast histopathology image dataset outperformed the other model by a substantial 13.2% accuracy, despite the difference in organ sites between the source and target datasets. Similarly, Ijaz et al [31] introduced a transfer learning approach by utilizing modified versions of VGG16 and VGG19 pretrained on source lung, colon, and breast cancer datasets, which were then applied to a target breast cancer dataset for binary classification. This approach achieved the highest accuracy of 98.96% on the x400 magnification factor subset, outperforming other models by margins ranging from 2.65% to 15.81% accuracy.…”
Section: B Intra-domain Transfer Learningmentioning
confidence: 99%
“… 13 , 14 , 15 Recent studies use attention-based models to construct slide-level representation by aggregating weighted tile-level features, 12 , 16 , 17 , 18 , 19 e.g., via multi-head attention (MHA), hierarchical attention, dual attention, or convolutional block attention modules. 12 , 16 , 20 , 21 , 22 , 23 , 24 Other important approaches have included multiscale attention and vision transformer models, which utilize the correlations across tiles to improve slide-level representations. 20 , 25 , 26 , 27 , 28 Fully CNN-based approaches have also been considered for attention-like functions, e.g., encoding of tiles by a CNN followed by an additional deep CNN, 29 possibly with multi-scale tiling.…”
Section: Introductionmentioning
confidence: 99%