2022
DOI: 10.48550/arxiv.2203.12175
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Adaptive Transformers for Robust Few-shot Cross-domain Face Anti-spoofing

Abstract: While recent face anti-spoofing methods perform well under the intra-domain setups, an effective approach needs to account for much larger appearance variations of images acquired in complex scenes with different sensors for robust performance. In this paper, we present adaptive vision transformers (ViT) for robust cross-domain face anti-spoofing. Specifically, we adopt ViT as a backbone to exploit its strength to account for long-range dependencies among pixels. We further introduce the ensemble adapters modu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 55 publications
0
2
0
Order By: Relevance
“…Tran-sRPPG [62] proposes a pure rPPG transformer framework for mining the global relationship within MSTmaps for liveness representation. ViTAF [63] uses the pure ViT to solve the zero-and few-shot, face anti-spoofing task. Contemporaneous, MA-ViT [64] aims to solve flexible modal face antispoofing by introducing Modality-Agnostic Transformer Block (MATB), which consists of two stacked attentions named Modal-Disentangle Attention (MDA) and Cross-Modal Attention (CMA), to eliminate modality-related information for each modal sequences and supplement modality-agnostic liveness features from another modal sequences, respectively.…”
Section: Related Workmentioning
confidence: 99%
“…Tran-sRPPG [62] proposes a pure rPPG transformer framework for mining the global relationship within MSTmaps for liveness representation. ViTAF [63] uses the pure ViT to solve the zero-and few-shot, face anti-spoofing task. Contemporaneous, MA-ViT [64] aims to solve flexible modal face antispoofing by introducing Modality-Agnostic Transformer Block (MATB), which consists of two stacked attentions named Modal-Disentangle Attention (MDA) and Cross-Modal Attention (CMA), to eliminate modality-related information for each modal sequences and supplement modality-agnostic liveness features from another modal sequences, respectively.…”
Section: Related Workmentioning
confidence: 99%
“…The method SSAN [40] is designed to split the representation into content and style ones with different supervision. ViTAF method [19] uses adaptive transformers as backbone and has good performance on cross-domain task. But this model requires large additional datasets for supervised pretraining.…”
Section: Introductionmentioning
confidence: 99%