2021
DOI: 10.36227/techrxiv.14912367.v1
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

xViTCOS: Explainable Vision Transformer Based COVID-19 Screening Using Radiography

Abstract: Since its outbreak, the rapid growth of COrona VIrus Disease 2019 (COVID-19) across the globe has pushed the health care system in many countries to the verge of collapse. Therefore, it is imperative to correctly identify COVID-19 positive patients and isolate them as soon as possible to contain the spread of the disease and reduce the ongoing burden on the healthcare system. The primary COVID-19 screening test, RT-PCR although accurate and reliable, has a long turn-around time. In the recent past, several res… Show more

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
9
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
2
2
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 9 publications
(9 citation statements)
references
References 0 publications
0
9
0
Order By: Relevance
“…The authors in [43] propose a convolutional CapsNet for COVID-19 detection from chest X-ray images in binary as well as multi-class classification settings. xViTCOS [44] propose a vision transformer based deep neural classifier for COVID-19 prognosis.…”
Section: Related Workmentioning
confidence: 99%
“…The authors in [43] propose a convolutional CapsNet for COVID-19 detection from chest X-ray images in binary as well as multi-class classification settings. xViTCOS [44] propose a vision transformer based deep neural classifier for COVID-19 prognosis.…”
Section: Related Workmentioning
confidence: 99%
“…They need a vast receptive eld to capture long-range dependencies, which means developing large kernels or highly massive networks, resulting in an extremely complex model that is challenging to train. [9]. To overcome these drawbacks of CNN, some researchers have used other architectures, such as Capsule Neural Networks (Capsnets) [24] and ViT [8], for COVID-19 classi cation, which differs from the traditional CNN networks.…”
Section: Related Workmentioning
confidence: 99%
“…The trained model performed better than CNN, such as E cientNet-B0, Inception-V3, and ResNet-50 in a multi-classi cation challenge, with 92% accuracy and 98% AUC. Mondal et al [9] (2) iCTCF [30], which comes from a total of 1521 patients in two hospitals of Huazhong University of Science and Technology, China, including 894 COVID-19 pneumonia cases (including mild, severe, and critical cases), 328 novel coronavirus-negative patients (control group), and 299 patients with suspected COVID-19; (3) COVID-CTSet [31], which comes from the dataset of Negin Medical Center in Sari, Iran, including 377 patients with con rmed COVID-19, 95 novel coronavirus-negative patients, and 282 other pneumonia patients; and the remaining were collected from (4) TCIA [32], (5) COVID-19 Infection Segmentation Dataset [33], (6) LIDC-IDRI [34], (7) Radiopaedia [35], (8) and MosMedData [36]. In Table 1, MUST-COVID-19 contains images of about three classes, with 80 percent of images employed for training and veri cation and 20 percent for model testing.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Krishnan et al [17] and Park et al [39] utilize ViT-based models to achieve higher COVID-19 classification accuracy through CXR images. COVID-Transformer [48] and xViTCOS [33] have been proposed to further improve classification accuracy and focus on diagnosis-related regions. However, there is still much room for improvement to train ViT models in a small dataset, such as medical imaging dataset.…”
Section: Vision Transformermentioning
confidence: 99%