2014
DOI: 10.1016/j.tcs.2013.09.027
|View full text |Cite
|
Sign up to set email alerts
|

Domain adaptation and sample bias correction theory and algorithm for regression

Abstract: We present a series of new theoretical, algorithmic, and empirical results for domain adaptation and sample bias correction in regression. We prove that the discrepancy is a distance for the squared loss when the hypothesis set is the reproducing kernel Hilbert space induced by a universal kernel such as the Gaussian kernel. We give new pointwise loss guarantees based on the discrepancy of the empirical source and target distributions for the general class of kernel-based regularization algorithms. These bound… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
112
0

Year Published

2015
2015
2022
2022

Publication Types

Select...
4
2
1

Relationship

1
6

Authors

Journals

citations
Cited by 118 publications
(130 citation statements)
references
References 18 publications
0
112
0
Order By: Relevance
“…The discrepancy has several advantages over a measure such as the L 1 or total variation distance (Cortes and Mohri, 2013): it is a finer measure than the L 1 distance, it takes into account the loss function and the hypothesis set, it can be accurately estimated from finite samples for common hypothesis sets such as kernel-based ones, it is symmetric and verifies the triangle inequality. It further defines a distance in the case of an L p loss used with a universal kernel such as a Gaussian kernel.…”
Section: Previous Workmentioning
confidence: 99%
See 3 more Smart Citations
“…The discrepancy has several advantages over a measure such as the L 1 or total variation distance (Cortes and Mohri, 2013): it is a finer measure than the L 1 distance, it takes into account the loss function and the hypothesis set, it can be accurately estimated from finite samples for common hypothesis sets such as kernel-based ones, it is symmetric and verifies the triangle inequality. It further defines a distance in the case of an L p loss used with a universal kernel such as a Gaussian kernel.…”
Section: Previous Workmentioning
confidence: 99%
“…Using q min instead of Q amounts to reweighting the loss on the training samples to minimize the discrepancy between the empirical distribution and P . Besides its theoretical motivation, this algorithm has been shown to outperform several other algorithms in a series of experiments carried out by (Cortes and Mohri, 2013). Observe that, by definition, the solution q min of discrepancy minimization is obtained by minimizing a maximum over all pairs of hypotheses, that is max h,h ∈H |L P (h, h ) − L q min (h, h )|.…”
Section: Previous Workmentioning
confidence: 99%
See 2 more Smart Citations
“…This behavior is generally shared by all supervised approaches to bio/geophysical parameter retrieval and suggests that, before applying the trained estimator to geographical regions distinct from those where the training samples were located, additional testing with data collected from those regions is expected to be necessary. The combination with domain adaptation techniques (e.g., [67]) could also be an interesting extension aimed at favoring application to areas without training samples.…”
Section: Discussionmentioning
confidence: 99%