Dataset Shift in Machine Learning 2008
DOI: 10.7551/mitpress/9780262170055.003.0008
|View full text |Cite
|
Sign up to set email alerts
|

Covariate Shift by Kernel Mean Matching

Abstract: Given sets of observations of training and test data, we consider the problem of re-weighting the training data such that its distribution more closely matches that of the test data. We achieve this goal by matching covariate distributions between training and test sets in a high dimensional feature space (specifically, a reproducing kernel Hilbert space). This approach does not require distribution estimation. Instead, the sample weights are obtained by a simple quadratic programming procedure. We provide a u… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
527
0

Year Published

2011
2011
2023
2023

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 465 publications
(654 citation statements)
references
References 26 publications
1
527
0
Order By: Relevance
“…. , X n without going through density estimation (Gretton et al 2009). The basic idea of KMM is to find w 0 (x) such that the mean discrepancy between nonlinearly transformed samples drawn from P and Q is minimized in a universal reproducing kernel Hilbert space (Steinwart 2001).…”
Section: Kernel Mean Matching (Kmm)mentioning
confidence: 99%
See 1 more Smart Citation
“…. , X n without going through density estimation (Gretton et al 2009). The basic idea of KMM is to find w 0 (x) such that the mean discrepancy between nonlinearly transformed samples drawn from P and Q is minimized in a universal reproducing kernel Hilbert space (Steinwart 2001).…”
Section: Kernel Mean Matching (Kmm)mentioning
confidence: 99%
“…The kernel mean matching (KMM) method (Gretton et al 2009) directly gives estimates of the density ratio by matching the two distributions using universal reproducing kernel Hilbert spaces (Steinwart 2001). KMM can be regarded as a kernelized variant of Qin's moment matching estimator (Qin 1998).…”
Section: Introductionmentioning
confidence: 99%
“…Thus, direct density-ratio estimation is substantially easier than density estimation . Following this idea, methods of direct density-ratio estimation have been developed , e.g., kernel mean matching (Gretton et al, 2009), the logistic-regression method (Bickel et al, 2007), and the Kullback-Leibler importance estimation procedure (KLIEP) (Sugiyama et al, 2008). In the context of change-point detection, KLIEP was reported to outperform other approaches (Kawahara and Sugiyama, 2012) such as the one-class support vector machine (Schölkopf et al, 2001;Desobry et al, 2005) and singular-spectrum analysis (Moskvina and Zhigljavsky, 2003b).…”
Section: Introductionmentioning
confidence: 99%
“…Then an additional variable, s i , for each sample of training data set is defined [21,14]. s i is set to depend only on one of the sample features, therefore, the biasing procedure is called, simple bias [6]. This additional variable determines whether the corresponding sample is contributing in the biased training data set or not.…”
Section: Real World Data Setsmentioning
confidence: 99%
“…"Covariate shift" [5,6] and "class imbalance" [7] are two examples with different initial assumptions:…”
Section: Overviewmentioning
confidence: 99%