2011
DOI: 10.1186/1687-6180-2011-62
|View full text |Cite
|
Sign up to set email alerts
|

Distance-based features in pattern classification

Abstract: In data mining and pattern classification, feature extraction and representation methods are a very important step since the extracted features have a direct and significant impact on the classification accuracy. In literature, numbers of novel feature extraction and representation methods have been proposed. However, many of them only focus on specific domain problems. In this article, we introduce a novel distance-based feature extraction method for various pattern classification problems. Specifically, two … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2016
2016
2023
2023

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 13 publications
(6 citation statements)
references
References 18 publications
(17 reference statements)
0
6
0
Order By: Relevance
“…However, other current trend not explored in this paper is the use of unsupervised learning to compute new features from existing ones. Some examples of this are the approach proposed by [40] for intrusion detection or the more general method presented in [41], that expand the original feature vectors by computing new features using distances from each data sample to a number of centroids found by a clustering algorithm. In this same line, the use of distance-based features to a set of reference patterns, or the related concept of pairwise dissimilarities [42], could effectively be used to enrich or replace the information provided in the original feature vectors.…”
Section: Discussionmentioning
confidence: 99%
“…However, other current trend not explored in this paper is the use of unsupervised learning to compute new features from existing ones. Some examples of this are the approach proposed by [40] for intrusion detection or the more general method presented in [41], that expand the original feature vectors by computing new features using distances from each data sample to a number of centroids found by a clustering algorithm. In this same line, the use of distance-based features to a set of reference patterns, or the related concept of pairwise dissimilarities [42], could effectively be used to enrich or replace the information provided in the original feature vectors.…”
Section: Discussionmentioning
confidence: 99%
“…The k-means algorithm is generally used to identify centroids (averaging object) in many studies on distance-based feature extraction [7], [43] since it is easy to realize and its computational cost is low. However, k-means has two drawbacks.…”
Section: ) Energy Normalizationmentioning
confidence: 99%
“…After extracting the wavelet ridge, we employ three preprocessing methods on the wavelet ridge: the elimination of the influence of noise using SVD denoising, which has a time complexity of (n 2 ), dimensionality reduction using PAA which has a time complexity of (m), where m denotes the reduced dimension, and normalization with respect to energy, which has a linear time complexity of (m). For the existing M -dimensional training data, the time complexity of the DBA algorithm is (I • M • m 2 ) [43], where the parameter I denotes the number of iterations. Obviously, when large training datasets or long signal lengths appear, time complexity will increase rapidly.…”
Section: E Computational Complexity Analysismentioning
confidence: 99%
“…For this purpose, the multicategory support vector machine (M-SVM) method can be used. For more details in this regard, one can refer to [13,22].…”
Section: Estimation Of Subsystems and Subregionsmentioning
confidence: 99%
“…Also in [21], a method using selforganizing map (SOM)-based spectral clustering is proposed for agriculture management. In [22], a novel distance-based feature extraction method for various pattern classification problems is introduced. Specifically, two distances are extracted, which are based on (1) the distance between the data and its intra-cluster center and (2) the distance between the data and its extracluster centers.…”
Section: Introductionmentioning
confidence: 99%