2023
DOI: 10.1016/j.jare.2022.08.021
|View full text |Cite
|
Sign up to set email alerts
|

A hybrid explainable ensemble transformer encoder for pneumonia identification from chest X-ray images

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
30
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
7
1
1

Relationship

3
6

Authors

Journals

citations
Cited by 61 publications
(40 citation statements)
references
References 65 publications
0
30
0
Order By: Relevance
“…In the future, the used models and possibly other ones can be investigated on mixed images collected from datasets that have different intensities, such as INbreast, DDSM, and MAIS datasets, helping to find the best models that can deal with breast cancer images with different densities. We have a plan to continue improving the performance behavior and providing more interesting breast cancer prediction results using the newly impressive AI technologies such as explainable AI [ 48 , 49 , 50 ], federated learning [ 51 ], and so on. It is known that the medical images always have common characteristics that contain similarities in contextual features, and any deep learning model should be retuned again with respect to each modality.…”
Section: Resultsmentioning
confidence: 99%
“…In the future, the used models and possibly other ones can be investigated on mixed images collected from datasets that have different intensities, such as INbreast, DDSM, and MAIS datasets, helping to find the best models that can deal with breast cancer images with different densities. We have a plan to continue improving the performance behavior and providing more interesting breast cancer prediction results using the newly impressive AI technologies such as explainable AI [ 48 , 49 , 50 ], federated learning [ 51 ], and so on. It is known that the medical images always have common characteristics that contain similarities in contextual features, and any deep learning model should be retuned again with respect to each modality.…”
Section: Resultsmentioning
confidence: 99%
“…For direct comparison using the same dataset, four deep learning models of DenseNet 201, ResNet50, Inception-V3, and Mobilenet-V2 are adopted and used. These AI models are selected to perform such direct comparison due to their promising classification performance in the research domain [ 21 , 37 , 47 , 58 , 59 , 64 , 65 ]. Such comparison is important to investigate the reliability of the proposed model with the trusted ones.…”
Section: Discussionmentioning
confidence: 99%
“…The evaluation results shown in the result section are achieved over a 5-fold cross-validation test to investigate the reliability and feasibility of the proposed BCNet. The definition of the evaluation metrics is summarized in Equations (1)–(7) [ 20 , 59 , 60 , 61 , 62 , 63 ]. True positive (TP), true negative (TN), false positive (FP), and false negative (FN) are derived via a multi-class confusion matrix for each fold test.…”
Section: Methodsmentioning
confidence: 99%
“…In comparison to state-of-the-art CNN in image classification evaluations, the Vision Transformer (ViT) performs well. It is one of the favourable attempts to exploit Transformer specifically on images (18; 45; 46). Despite having better performance, it has an easy-to-use modular framework that allows for wide-ranging application in multiple tasks with minimal modification.…”
Section: Related Workmentioning
confidence: 99%