2011
DOI: 10.1111/j.1467-9868.2011.01003.x
|View full text |Cite
|
Sign up to set email alerts
|

Achieving near Perfect Classification for Functional Data

Abstract: Summary.  We show that, in functional data classification problems, perfect asymptotic classification is often possible, making use of the intrinsic very high dimensional nature of functional data. This performance is often achieved by linear methods, which are optimal in important cases. These results point to a marked contrast between classification for functional data and its counterpart in conventional multivariate analysis, where the dimension is kept fixed as the sample size diverges. In the latter setti… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

5
219
0

Year Published

2015
2015
2022
2022

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 139 publications
(224 citation statements)
references
References 32 publications
5
219
0
Order By: Relevance
“…Methods of functional data analysis are becoming increasingly popular, e.g. in the cluster analysis (Jacques and Preda 2013;James and Sugar 2003;Peng and Müller 2008), classification (Chamroukhi et al 2013;Delaigle and Hall 2012;Mosler and Mozharovskyi 2015;Rossi and Villa 2006) and regression (Ferraty et al 2012;Goia and Vieu 2014;Kudraszow and Vieu 2013;Peng et al 2015;Rachdi and Vieu 2006;Wang et al 2015). Unfortunately, multivariate data methods cannot be directly used for functional data, because of the problem of dimensionality and difficulty in putting functional data into order.…”
Section: Introductionmentioning
confidence: 99%
“…Methods of functional data analysis are becoming increasingly popular, e.g. in the cluster analysis (Jacques and Preda 2013;James and Sugar 2003;Peng and Müller 2008), classification (Chamroukhi et al 2013;Delaigle and Hall 2012;Mosler and Mozharovskyi 2015;Rossi and Villa 2006) and regression (Ferraty et al 2012;Goia and Vieu 2014;Kudraszow and Vieu 2013;Peng et al 2015;Rachdi and Vieu 2006;Wang et al 2015). Unfortunately, multivariate data methods cannot be directly used for functional data, because of the problem of dimensionality and difficulty in putting functional data into order.…”
Section: Introductionmentioning
confidence: 99%
“…Note that any functional distance can be used to implement the procedure. In particular, Delaigle and Hall (2012) considered the case of G = 2 classes that have different mean and a common covariance operator and proposed to project the functions into a given direction and then compute the squared Euclidean distance between the observations. More precisely, Delaigle and Hall (2012) proposed to use the centroid classifier with the distance between χ 0 and the sample functional mean µ χg , for g = 1, 2, denoted by DH, and given by:…”
Section: The Centroid Proceduresmentioning
confidence: 99%
“…As it can be seen in the figure, the four scenarios appear to be complicated scenarios for classification purposes. For each generated dataset, the functional observations in the test sample are classified using the following procedures: (1) the kNN procedure with seven different functional distances, the L 1 , L 2 and L ∞ distances as proposed by Baíllo et al (2011), the functional principal components (FPC) semi-distance assuming either a common or a different covariance operator, denoted by F P C C and F P C D , respectively, and the functional Mahalanobis (FM) semi-distance assuming either a common or a different covariance operator, denoted by F M C and F M D , respectively, as proposed in Section 3; (2) the centroid procedure with eight different functional distances, the first seven as in the kNN procedure and the distance proposed by Delaigle and Hall (2012) given in (17) and denoted by DH; (3) the linear and quadratic Bayes classification rules as proposed in Section 3, denoted by F LBCR and F QBCR, respectively; and (4) the multivariate linear and quadratic Bayes classification rules applied on the coefficients of the B-splines basis representation, denoted by LBCR Coef. and QBCR Coef., respectively.…”
Section: Monte Carlo Studymentioning
confidence: 99%
See 2 more Smart Citations