Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery &Amp; Data Mining 2018
DOI: 10.1145/3219819.3220082
|View full text |Cite
|
Sign up to set email alerts
|

Stable Prediction across Unknown Environments

Abstract: In many machine learning applications, the training distribution used to learn a probabilistic classifier differs from the testing distribution on which the classifier will be used to make predictions. Traditional methods correct the distribution shift by reweighting the training data with the ratio of the density between test and training data. But in many applications training takes place without prior knowledge of the testing. Recently, methods have been proposed to address the shift by learning causal stru… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
116
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
2
1

Relationship

2
6

Authors

Journals

citations
Cited by 126 publications
(117 citation statements)
references
References 21 publications
1
116
0
Order By: Relevance
“…A qualified model for student performance prediction should have good results from both regression and classification perspectives. In this paper, we evaluated the prediction performance of all models using four widely-used metrics in the domain [13], [24], [49], [50], [56]. From the regression perspective, we selected Mean Absolute Error (MAE) and Root Mean Square Error (RMSE), to quantify the distance between predicted scores and the actual ones.…”
Section: Evaluation Metricsmentioning
confidence: 99%
“…A qualified model for student performance prediction should have good results from both regression and classification perspectives. In this paper, we evaluated the prediction performance of all models using four widely-used metrics in the domain [13], [24], [49], [50], [56]. From the regression perspective, we selected Mean Absolute Error (MAE) and Root Mean Square Error (RMSE), to quantify the distance between predicted scores and the actual ones.…”
Section: Evaluation Metricsmentioning
confidence: 99%
“…The joint distribution of features and outcomes on (X, Y ) can change across environments: P e XY = P e XY for e, e ∈ E. In this paper, our goal is to learn a predictive model for stable prediction with model misspecification and agnostic distribution shift. To measure its performance on stable prediction problem, we adopt the Average Error and Stability Error in (Kuang et al 2018) as:…”
Section: Stable Prediction Problemmentioning
confidence: 99%
“…where |E| refers to the number of test environments, and RM SE(D e ) represents the Root Mean Square Error of a predictive model on dataset D e . Actually, Average Error and Stability Error refer to the mean and variance of predictive error over all possible environments e ∈ E. Then, the stable prediction problem (Kuang et al 2018) is defined as: Problem 1 (Stable Prediction) Given one training environment e ∈ E with dataset D e = (X e , Y e ), the task is to learn a predictive model to predict across unknown environment E with not only small Average Error but also small Stability Error.…”
Section: Stable Prediction Problemmentioning
confidence: 99%
See 1 more Smart Citation
“…A fundamental requirement for out-of-domain transfer learning is to mitigate the biases from the pretraining data [49], which may be useful for the in-domain testing but harmful for out-of-domain testing [19] due to the spurious correlation [34]. To verify such existence of the correlation biases, we follow [49] to conduct a toy experiment on Conceptual Caption dataset.…”
Section: Introductionmentioning
confidence: 99%