2022
DOI: 10.1007/s10489-022-03867-9
|View full text |Cite
|
Sign up to set email alerts
|

DeepFake detection algorithm based on improved vision transformer

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
1
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
5

Relationship

0
10

Authors

Journals

citations
Cited by 42 publications
(10 citation statements)
references
References 26 publications
0
1
0
Order By: Relevance
“…Arshed et al [123] applied vision transformers over deepfakes and got excellent results on deepfake images shared over Kaggle. Similar results were observed by Heo et al [124] with deepfake videos. They combine patch-based positioning and vector-concatenated CNN features to interact with all positions to determine the artifact region.…”
Section: Convolutional Neural Network-image Specificsupporting
confidence: 91%
“…Arshed et al [123] applied vision transformers over deepfakes and got excellent results on deepfake images shared over Kaggle. Similar results were observed by Heo et al [124] with deepfake videos. They combine patch-based positioning and vector-concatenated CNN features to interact with all positions to determine the artifact region.…”
Section: Convolutional Neural Network-image Specificsupporting
confidence: 91%
“…Its remarkable capability to capture extensive long-range and global contextual information has stimulated the adaptation of transformer to computer vision (CV) tasks [15,32] and derived a series of architectures called vision transformers (ViTs). For face forgery detection, some works have [17,33] directly reshaped the features extracted by CNN into a series of low-dimensional patches and passed them to VITs encoder, achieving certain levels of generalization performance. M2TR [34] operated patches of different sizes to model a multiscale transformer.…”
Section: 2mentioning
confidence: 99%
“…Moreover, compared to state-of-the-art solutions, the method achieved the highest mean accuracy in the four face manipulation strategies of the FaceForensics++ dataset. (Heo et al, 2023) proposed a DeepFake detection using a Vision Transformer Model, which has indicated good performance in recent image classifications and combined CNN and patch-embedding features during the input stage. The Robust Vision Transformer Model has shown efficiency compared with EfficientNet as the state-of-the-art model, which consists of a 2D CNN network.…”
Section: Cnn+grumentioning
confidence: 99%