Perspectives of Systems Informatics
DOI: 10.1007/978-3-540-70881-0_24
|View full text |Cite
|
Sign up to set email alerts
|

On the Importance of Parameter Tuning in Text Categorization

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
7
0

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 11 publications
(7 citation statements)
references
References 11 publications
0
7
0
Order By: Relevance
“…For the text‐categorization domain, different settings are used, such as β = 16, γ = 4, and β = γ, where Moschitti () showed that β = γ is the best literature parameterization. However, Koster and Beney () found that the β parameter must be smaller (or even much smaller) than the γ parameter, in which β = 0.1 and γ = 1 were used in their study, so that a greater weight is given to negative examples.…”
Section: Important Factors Of Prfmentioning
confidence: 99%
“…For the text‐categorization domain, different settings are used, such as β = 16, γ = 4, and β = γ, where Moschitti () showed that β = γ is the best literature parameterization. However, Koster and Beney () found that the β parameter must be smaller (or even much smaller) than the γ parameter, in which β = 0.1 and γ = 1 were used in their study, so that a greater weight is given to negative examples.…”
Section: Important Factors Of Prfmentioning
confidence: 99%
“…In related text retrieval studies, different settings are used. For example, Moschitti [20] concluded that β ¼γ is the best literature parameterization whereas Koster and Beney [16] found that β must be smaller than γ to obtain better retrieval performance. These findings indicate that the best parameter setting should be dataset dependent.…”
Section: Further Comparisonsmentioning
confidence: 98%
“…Rocchio's (1971) method was one of the most famous and successful methods. Also, lots of variations with the similar idea had achieved good performances: Ide (1971) used the top irrelevant document only for feedback; Singhal et al (1997) indicated that better results could be obtained by using documents close to the query of interest only, rather than all documents; Desjardins and Godin (2000) presented the development of a different weighting formula; Nick and Themis' (2001) approach was to rank the rating of the relevance of the retrieved documents; Kim et al's (2001) study calculated the relevance degree; Azimi-Sadjadi et al (2007) study exploited relevance feedback from multiple expert users; Koster and Beney (2007) proposed the modification of parameters in Rocchio's original formula.…”
Section: The Application Of Relevance Feedbackmentioning
confidence: 99%