1986
DOI: 10.1214/ss/1177013622
|View full text |Cite
|
Sign up to set email alerts
|

Influential Observations, High Leverage Points, and Outliers in Linear Regression

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

4
392
0
4

Year Published

1988
1988
2017
2017

Publication Types

Select...
4
4

Relationship

0
8

Authors

Journals

citations
Cited by 655 publications
(420 citation statements)
references
References 25 publications
4
392
0
4
Order By: Relevance
“…3 are, up to scaling, equal to the diagonal elements of the so-called "hat matrix," i.e., the projection matrix onto the span of the top k right singular vectors of A (19,20). As such, they have a natural statistical interpretation as a "leverage score" or "influence score" associated with each of the data points (19)(20)(21). In particular, π j quantifies the amount of leverage or influence exerted by the jth column of A on its optimal low-rank approximation.…”
Section: Statistical Leverage and Improved Matrix Decompositionsmentioning
confidence: 99%
“…3 are, up to scaling, equal to the diagonal elements of the so-called "hat matrix," i.e., the projection matrix onto the span of the top k right singular vectors of A (19,20). As such, they have a natural statistical interpretation as a "leverage score" or "influence score" associated with each of the data points (19)(20)(21). In particular, π j quantifies the amount of leverage or influence exerted by the jth column of A on its optimal low-rank approximation.…”
Section: Statistical Leverage and Improved Matrix Decompositionsmentioning
confidence: 99%
“…Concepts used in regression diagnostics such as influential observations, leverage points and hat matrices are solely based on the design matrix A (see e.g. Belsley et al 1980;Chatterjee and Hadi 1986;Barnett and Lewis 1994). Fortunately, the need of practical application of these concepts has led geodesists to incorporate the weight matrix into the measures related to these concepts (see e.g.…”
Section: The Effect Of Weights Of Observations On Robustnessmentioning
confidence: 99%
“…Identification and removal of contaminated data have been attempted and realized in two different ways, either (1) by first cleaning the data and then applying the classical least squares criterion to the remaining data (see e.g. Anscombe 1960;Baarda 1968;Pope 1976;Belsley et al 1980;Chatterjee and Hadi 1986;Barnett and Lewis 1994); or (2) by designing robust estimation criteria and applying them directly to contaminated data (see e.g. Huber 1981;Hampel et al 1986;Jurecková and Sen 1996;Koch 1999).…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Feature extraction and dimension reduction can be combined in one step by using ''multiple linear regression'' model which performs a mapping of the multi-regime data to a lower-dimensional space in such a way that the variance of the measurements in the low-dimensional finding is maximized. Multiple linear regression calculates the relationship between different explanatory variables and a target variable by fitting a linear equation to observed data [15,16]. This model is based on:…”
Section: Signal Processing and Dimensionality Reductionmentioning
confidence: 99%