2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2022
DOI: 10.1109/cvpr52688.2022.00969
|View full text |Cite
|
Sign up to set email alerts
|

DAFormer: Improving Network Architectures and Training Strategies for Domain-Adaptive Semantic Segmentation

Abstract: As acquiring pixel-wise annotations of real-world images for semantic segmentation is a costly process, a model can instead be trained with more accessible synthetic data and adapted to real images without requiring their annotations. This process is studied in unsupervised domain adaptation (UDA). Even though a large number of methods propose new adaptation strategies, they are mostly based on outdated network architectures. As the influence of recent network architectures has not been systematically studied,… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
174
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
5
2

Relationship

1
6

Authors

Journals

citations
Cited by 256 publications
(241 citation statements)
references
References 79 publications
1
174
0
Order By: Relevance
“…Adversarial training follows a GAN framework [23,25] to aligns the source and target domains feature distributions at input [24,28], output [77,79], patch [13], or feature level [29,77]. In self-training, the supervision for target domain comes from pseudo-labels [38] which can be computed offline [65,86,93,94] or online [76,80,31]. Consistency regularization [70,75] or label prototypes [88] formulated on CDMS [76,91] or data augmentation [3,14,49] DA action recognition and detection.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Adversarial training follows a GAN framework [23,25] to aligns the source and target domains feature distributions at input [24,28], output [77,79], patch [13], or feature level [29,77]. In self-training, the supervision for target domain comes from pseudo-labels [38] which can be computed offline [65,86,93,94] or online [76,80,31]. Consistency regularization [70,75] or label prototypes [88] formulated on CDMS [76,91] or data augmentation [3,14,49] DA action recognition and detection.…”
Section: Related Workmentioning
confidence: 99%
“…We don't consider multi-label action classes, i.e., those action categories in which action instances could have more than one class label. A rare-class sampling [31] could be helpful to generate more diversified training samples for the under-represented classes. precision of class take a photo improves over 10%.…”
Section: Comparison To State-of-the-artmentioning
confidence: 99%
“…By leveraging the model itself to generate pseudo-labels on unlabeled data, self-training together with tailored strategies such as consistency regularization [77,3], cross-domain mixup [55,78], contrastive learning [31,41,79,36], pseudo-label refine [63,73,76], auxiliary tasks [60,62] and class balanced training [35] achieves excellent performance. Recently, Hoyer et al [29] empirically proved that the transformer architecture [67] is more robust to domain shift than CNN.…”
Section: Related Workmentioning
confidence: 99%
“…Improving the labeling efficiency of deep learning algorithms is vital in practice since acquiring high-quality annotations could consume great effort. Self-training (ST) offers a promising solution to alleviate this issue by learning with limited labeled data and large-scale unlabeled data [57,29]. The key thought is to learn a model on labeled samples and use it to generate pseudo-labels for unlabeled samples to teach the model itself.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation