2010 Fifth International Conference on Digital Information Management (ICDIM) 2010
DOI: 10.1109/icdim.2010.5664669
|View full text |Cite
|
Sign up to set email alerts
|

Latent semantic indexing and large dataset: Study of term-weighting schemes

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2011
2011
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 9 publications
(6 citation statements)
references
References 7 publications
0
6
0
Order By: Relevance
“…Three different term weighting schemes and own list of stop words have used to judge the performance. Recall, Precision and Coefficient of Variation were used to evaluate the retrieval performance of LSI based retrieval system [17]. Khaled M. Hammouda explored the concept of document indexing technique with more informative features including phrases and their weights that are important for indexing.…”
Section: Related Workmentioning
confidence: 99%
“…Three different term weighting schemes and own list of stop words have used to judge the performance. Recall, Precision and Coefficient of Variation were used to evaluate the retrieval performance of LSI based retrieval system [17]. Khaled M. Hammouda explored the concept of document indexing technique with more informative features including phrases and their weights that are important for indexing.…”
Section: Related Workmentioning
confidence: 99%
“…However, traditional log entropy weighting takes an insufficient approach and fails to account for local and global attributes of words present within a corpus [11]- [13]. To address these limitations of log entropy weighting approaches such as log entropy weighting in this research, we developed TWLE as an improved weighting alternative that considers both local and global attributes of words [14], [15]. Local weights help capture context significance through evaluation of document frequency while global weights indicate importance within topic formation by considering corpus wide frequency [16], [17].…”
Section: Introductionmentioning
confidence: 99%
“…They are not considered influential during the execution of LSI process to retrieve relevant documents. It also reduces the size of the indexing structure considerably [9].…”
Section: Stop Word Removalmentioning
confidence: 99%
“…[9]. After building the term-document matrix Normalize it using where A is the TDM and n is the total no of words in document j.…”
Section: Term-document Matrixmentioning
confidence: 99%