Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Langua 2021
DOI: 10.18653/v1/2021.naacl-main.147
|View full text |Cite
|
Sign up to set email alerts
|

Domain Divergences: A Survey and Empirical Analysis

Abstract: Domain divergence plays a significant role in estimating the performance of a model in new domains. While there is a significant literature on divergence measures, researchers find it hard to choose an appropriate divergence for a given NLP application. We address this shortcoming by both surveying the literature and through an empirical study. We develop a taxonomy of divergence measures consisting of three classes -Information-theoretic, Geometric, and Higher-order measures and identify the relationships bet… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
8
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
7
2

Relationship

0
9

Authors

Journals

citations
Cited by 11 publications
(8 citation statements)
references
References 74 publications
(97 reference statements)
0
8
0
Order By: Relevance
“…Based on this distribution-difference-based bound, we can minimize the following objective in our model finetuning [36], to project representations of samples from different domains close together in the representation space: regularization for minimizing representation difference (5) One practical divergence measure for Δ(.) is the Central Moment Discrepancy [67], which is theoretically grounded, efficient to implement and compute, and has shown superior empirical success in learning domain invariant representations [44,74].…”
Section: Fine-tuning With Cmd-based Regularizationmentioning
confidence: 99%
“…Based on this distribution-difference-based bound, we can minimize the following objective in our model finetuning [36], to project representations of samples from different domains close together in the representation space: regularization for minimizing representation difference (5) One practical divergence measure for Δ(.) is the Central Moment Discrepancy [67], which is theoretically grounded, efficient to implement and compute, and has shown superior empirical success in learning domain invariant representations [44,74].…”
Section: Fine-tuning With Cmd-based Regularizationmentioning
confidence: 99%
“…Even in lower-stake scenarios a model trained on only a subset of the global data distribution can have inconsistent behaviour when applied to a different target data distribution (D'Amour et al, 2020;Koh et al, 2020). For instance, language or even domain differences can have a noticeable impact on model performance (White & Cotterell, 2021;Ramesh Kashyap et al, 2021). Increased data diversity can improve the generalization ability of models to new domains, e.g., more languages in NLP (Benjamin, 2018).…”
Section: Choice Of Datasetmentioning
confidence: 99%
“…The most optimal transferability metric would be resource-efficient, such that it is capable of accurately predicting the final performance of the model whilst minimising the amount of processing required to compute it. To this end, several works (Van Asch and Daelemans, 2010;Ruder and Plank, 2017;Ramesh Kashyap et al, 2021) have focused on estimating transferability prior to finetuning, using statistical measures of divergence between the underlying feature spaces of model pairs. Domain divergence measures are used to produce a notion of distance between pairs of domains by comparing their representations and have seen significant usage in works which investigate the correlation between their estimations and performance change (Van Asch and Daelemans, 2010;Ramesh Kashyap et al, 2021).…”
Section: Introductionmentioning
confidence: 99%