2006
DOI: 10.1016/j.ipm.2004.11.007
|View full text |Cite
|
Sign up to set email alerts
|

A framework for understanding Latent Semantic Indexing (LSI) performance

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
81
0
5

Year Published

2007
2007
2014
2014

Publication Types

Select...
7
2

Relationship

1
8

Authors

Journals

citations
Cited by 148 publications
(90 citation statements)
references
References 15 publications
1
81
0
5
Order By: Relevance
“…In LSA, researchers may use more than one weighting scheme, and may subject the results calculated to further factor analysis or clustering. We refer the reader to Kontostathis & Pottenger (2006) In conclusion, our work illustrates in a 'proof-ofconcept' approach how two modern computational analyses can be used, in isolation as well as in complementary fashion, to aid the content analysis of large corpi of text. We show how the results of one analysis (LSA) can be used to inform our understanding of the trends between separate data sets, and we demonstrate how a text mining analysis (using Leximancer) can be used to provide further insights for the underlying rationale of the outcomes of the LSA analysis.…”
Section: Discussionmentioning
confidence: 99%
“…In LSA, researchers may use more than one weighting scheme, and may subject the results calculated to further factor analysis or clustering. We refer the reader to Kontostathis & Pottenger (2006) In conclusion, our work illustrates in a 'proof-ofconcept' approach how two modern computational analyses can be used, in isolation as well as in complementary fashion, to aid the content analysis of large corpi of text. We show how the results of one analysis (LSA) can be used to inform our understanding of the trends between separate data sets, and we demonstrate how a text mining analysis (using Leximancer) can be used to provide further insights for the underlying rationale of the outcomes of the LSA analysis.…”
Section: Discussionmentioning
confidence: 99%
“…We determine the sparsity pattern of M 1 column by column. First, find the largest entry in each column of A, suppose they are a 31 , a 12 …”
Section: Multistep Approachmentioning
confidence: 99%
“…8 and it shows that our MSSP is not comparable with CPM. Compared with the methods used in [13,12,14,3], the size of NPL preprocessed in our way is much larger and the data matrix has the feature of m < n, i.e. the document's number is more than the term's number.…”
Section: Numerical Experimentsmentioning
confidence: 99%
“…where, T is a left singular vector representing a term by dimension matrix, S is a singular value dimension by dimension matrix and D is a right singular vector representing document by document matrix [31]. The decomposed matrices are then truncated to a dimension less than the original k-value and the original X matrix approximated in the reduced latent space which better represents semantic relationships between terms compared to the original k-dimension document space.…”
Section: Singular Value Decompositionmentioning
confidence: 99%