2017
DOI: 10.1109/tnnls.2016.2562670
|View full text |Cite
|
Sign up to set email alerts
|

Semisupervised Feature Selection Based on Relevance and Redundancy Criteria

Abstract: Feature selection aims to gain relevant features for improved classification performance and remove redundant features for reduced computational cost. How to balance these two factors is a problem especially when the categorical labels are costly to obtain. In this paper, we address this problem using semisupervised learning method and propose a max-relevance and min-redundancy criterion based on Pearson's correlation (RRPC) coefficient. This new method uses the incremental search technique to select optimal f… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
55
0

Year Published

2017
2017
2022
2022

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 134 publications
(55 citation statements)
references
References 47 publications
0
55
0
Order By: Relevance
“…Figure 1a shows the data set in the first two dimensions, in which two small Gaussian clusters are buried in one red class. We compared SADA with five methods, including sSelect (Zhao and Liu 2007), LSDF (Zhao, Lu, and H 2008), PRPC (Xu et al 2016), RLSR (Chen et al 2017) and DSFFS (Yuan et al 2018). In this experiment, the projection dimension was set as 1 and the nearest neighborhoods k was set as 5.…”
Section: Experimental Results and Analysismentioning
confidence: 99%
“…Figure 1a shows the data set in the first two dimensions, in which two small Gaussian clusters are buried in one red class. We compared SADA with five methods, including sSelect (Zhao and Liu 2007), LSDF (Zhao, Lu, and H 2008), PRPC (Xu et al 2016), RLSR (Chen et al 2017) and DSFFS (Yuan et al 2018). In this experiment, the projection dimension was set as 1 and the nearest neighborhoods k was set as 5.…”
Section: Experimental Results and Analysismentioning
confidence: 99%
“…As a filter feature selection method, Pearson correlation coefficient has strong generality and low complexity, and it has strong advantages in dealing with large-scale data sets and can eliminate a large number of irrelevant features in a short time. Therefore, it is often used for feature selection of the whole data set [34]. The formulation of Pearson correlation coefficient is represented in Equation 6:…”
Section: Proposed Emd-gru Model With Feature Selectionmentioning
confidence: 99%
“…Feature selection aims to gain relevant features and remove redundant features for reducing computational cost,() which can effectively avoid traveling all attributes. The max‐relevance and min‐redundancy criterion based on Pearson's correlation (RRPC) coefficient is an effective method to select optimal feature subsets. The RRPC algorithm achieves a good balance between relevance and redundancy in feature selection.…”
Section: A Fast Rank Mutual Information Based Decision Treementioning
confidence: 99%
“…Suppose the training data X with n condition attributes A=false{Akfalse}k=1n and one decision attribute D can be split into m subsets false{Xifalse}i=1m in the framework of Map‐Reduce. In the Map phase, we first apply RRPC algorithm to generate a subset A ′ of A from X i . Then, we compute the best splitting point cpjfalse(ifalse) for each attribute Ak in A ′ by the following formula: cpjfalse(ifalse)=argmaxcpjRMIcpjfalse(Akfalse), where RMIcpjfalse(Akfalse)=RMIfalse(Ak,cpj,Dfalse) is calculated by Equation and c p j is a center of A k generated by FCM algorithm.…”
Section: The Parallel Fast Rank Mutual Information Based Decision Treementioning
confidence: 99%