1996
DOI: 10.1002/(sici)1099-1506(199607/08)3:4<301::aid-nla84>3.0.co;2-s
|View full text |Cite
|
Sign up to set email alerts
|

Low-rank Orthogonal Decompositions for Information Retrieval Applications

Abstract: Current methods to index and retrieve documents from databases usually depend on a lexical match between query terms and keywords extracted from documents in a database. These methods can produce incomplete or irrelevant results due to the use of synonyms and polysemus words. The association of terms with documents (or implicit semantic structure) can be derived using large sparse {\it term‐by‐document} matrices. In fact, both terms and documents can be matched with user queries using representations in k‐spac… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
25
0

Year Published

1997
1997
2012
2012

Publication Types

Select...
4
3
2

Relationship

1
8

Authors

Journals

citations
Cited by 51 publications
(25 citation statements)
references
References 12 publications
0
25
0
Order By: Relevance
“…The sparse SVD function SVDS is based on the Arnoldi methods described in [36]. Note that, for practical purposes, less expensive factorizations such as QR or ULV may suffice in place of the SVD [10].…”
Section: Sparsitymentioning
confidence: 99%
“…The sparse SVD function SVDS is based on the Arnoldi methods described in [36]. Note that, for practical purposes, less expensive factorizations such as QR or ULV may suffice in place of the SVD [10].…”
Section: Sparsitymentioning
confidence: 99%
“…Compare the above expression with (3). Choosing the function φ(x) appropriately will allow us to interpretate this approach as a compromise between the vector space and the TSVD approaches.…”
Section: Latent Semantic Indexing By Polynomial Filteringmentioning
confidence: 99%
“…Most of the current implementations of LSI rely on matrix decompositions (see e.g., [3], [13]), with the truncated SVD (TSVD) being the most popular [1], [2]. In TSVD it is assumed that the smallest singular triplets are noisy and therefore only the largest singular triplets are used for the rank-k representation of the term-by-document matrix A.…”
Section: Introductionmentioning
confidence: 99%
“…These approximations identify hidden structures in word usage, thus enabling searches that go beyond simple keyword matching (see, for example [14]). Rank reduction techniques (SDD [40], SVD [11,12,47,10]) for type clustering are applicable here as they have been shown to be especially appropriate for latent semantic indexing in information retrieval. 1224 1288 1292 3 8 17 291 1386 1387 314 322 324 329 351 353 356 362 469 379 480 481 479 483 562 583 591 590 14 13 15 593 598 602 2437 605 2438 618 620 621 624 626 627 638 2486 652 688 691 719 722 729 734 2487 2488 742 2691 745 740 741 747 749 755 809 914 916 754 945 941 948 943 946 940 939 944 947 949 942 938 759 965 960 953 954 968 962 952 957 967 955 961 951 956 966 959 964 963 958 229 970 1008 1119 1127 1130 117 185 3797 3798 1157 1160 3799 1162 1167 1180 1182 1136 1011 1010 1017 1045 3903 1186 1201 1220 228 ...…”
Section: Clustering and Topical Compressionmentioning
confidence: 99%