2021
DOI: 10.48550/arxiv.2104.07235
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Vision Transformer using Low-level Chest X-ray Feature Corpus for COVID-19 Diagnosis and Severity Quantification

Abstract: Developing a robust algorithm to diagnose and quantify the severity of COVID-19 using Chest X-ray (CXR) requires a large number of well-curated COVID-19 datasets, which is difficult to collect under the global COVID-19 pandemic. On the other hand, CXR data with other findings are abundant. This situation is ideally suited for the Vision Transformer (ViT) architecture, where a lot of unlabeled data can be used through structural modeling by the self-attention mechanism. However, the use of existing ViT is not o… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
1
1

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(3 citation statements)
references
References 39 publications
0
3
0
Order By: Relevance
“…They used an encoder-decoder design. ViT was recently used to diagnose and predict the severity of COVID-19, demonstrating its SOTA performance [24].…”
Section: Vision Transformer Modelsmentioning
confidence: 99%
“…They used an encoder-decoder design. ViT was recently used to diagnose and predict the severity of COVID-19, demonstrating its SOTA performance [24].…”
Section: Vision Transformer Modelsmentioning
confidence: 99%
“…Recently, ViT was successfully used for diagnosis and severity prediction of COVID-19, showing the SOTA performance [43]. Specifically, to alleviate the overfitting problem with limited data available, the overall framework is decomposed into two steps: the pre-trained backbone network to classify common low-level CXR features, which was leveraged in the second step by Transformer for high-level diagnosis and severity prediction of COVID-19.…”
Section: Related Workmentioning
confidence: 99%
“…As suggested in Park et al [43], the head for classification was first initialized with pre-trained weights from the CheXpert dataset. We minimized the cross-entropy loss for the classification task.…”
Section: Implementation Detailsmentioning
confidence: 99%