2011
DOI: 10.18517/ijaseit.1.3.57
|View full text |Cite
|
Sign up to set email alerts
|

An Approach for Optimal Feature Subset Selection using a New Term Weighting Scheme and Mutual Information

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2012
2012
2021
2021

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 10 publications
(6 citation statements)
references
References 7 publications
0
6
0
Order By: Relevance
“…Fig. 3 Representation of SVM method [23] The second classifier is Naïve Bayes (NB), which performs well in text classification [24]. NB is a probabilistic classifier that considers the existence of the specific feature in class is independent of any other existence feature [14].…”
Section: Classificationmentioning
confidence: 99%
“…Fig. 3 Representation of SVM method [23] The second classifier is Naïve Bayes (NB), which performs well in text classification [24]. NB is a probabilistic classifier that considers the existence of the specific feature in class is independent of any other existence feature [14].…”
Section: Classificationmentioning
confidence: 99%
“…However, the major drawbacks of behavior-based are a considerable false positive rate (FP) and excessive monitoring time [14]. Further, the reduction of thousands of extracted features, evaluate similarities between them, and monitoring malware activities are directly effecting the ability of detecting zero-day malware attacks [17], [18].…”
Section: ) Heuristic-based Detectionmentioning
confidence: 99%
“…Several techniques have been developed to identify near-duplicate documents [610], web page duplicates [11–16], duplicate database records [17, 18] and bibliographic metadata [19]. Brin et al proposed the COPS (Copy Protection System) to protect important and intellectual property of original digital documents via registration of those documents on the system [6].…”
Section: Literature Reviewmentioning
confidence: 99%
“…Das et al proposed a TDW matrix-based algorithm with three phases: rendering, filtering and verification. In detail, receiving an input web page and a threshold in its first phase, prefix filtering and positional filtering to reduce the size of the record set in the second phase, and returning an optimal set of near-duplicate web pages in the verification (third) phase using the Minimum Weight Overlapping method [15] are the three phases of the algorithm.…”
Section: Literature Reviewmentioning
confidence: 99%