Proceedings of the Seventeenth Annual ACM Conference on Computer Science : Computing Trends in the 1990's Computing Trends in T 1989
DOI: 10.1145/75427.75440
|View full text |Cite
|
Sign up to set email alerts
|

A more cost effective algorithm for finding perfect hash functions

Abstract: of these contexts there are large static collections of records which must be indexed by keys, usually English words or phrases. Hashing is one method that can provide the desired rapid access, with little overhead in space, but unless the hash function chosen is suitable, there can be a considerable loss in performance due to collisions. Given adequate space, dynamic hashing methods have been developed to deal with this problem in changeable collections [ENBO88]. With static data sets, however, it is possible… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
8
0

Year Published

1989
1989
1994
1994

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(8 citation statements)
references
References 7 publications
0
8
0
Order By: Relevance
“…is of concern to us only insofar that good pseudo-random functions can be found which map the keys to integers. Several authors have given families of pseudo-random functions which work well for character string keys [4,5,6,7,10].…”
Section: Preliminariesmentioning
confidence: 99%
“…is of concern to us only insofar that good pseudo-random functions can be found which map the keys to integers. Several authors have given families of pseudo-random functions which work well for character string keys [4,5,6,7,10].…”
Section: Preliminariesmentioning
confidence: 99%
“…2 would work well as a perfect hashing algorithm because it completely eliminates the pattem collision problem, and also because it produces minimal word lists that can be ordered in any way (the list shown is alphabetical). If the trie array were stored in memory, the CPU time needed to index a word through the hie would actually be less than that of other perfect hashing functions currently in use for handling large word sets [22], [9]. For example, to look up the word CORN using Sager's algorithm [22] (or Fox's improvement of it [9]) requires that seven characters be indexed from the word, along with 11 additions, three MOD operations, and three array indexings.…”
Section: The N I E Data Structurementioning
confidence: 99%
“…CD-ROM's also have slow track seek times, which makes multiple accesses to the disk extremely frustrating for the user of the CD-ROM. Perfect hashing functions have been used in this environment to provide quick access to large medical dictionaries [9] and other word lists. The perfect hashing function allows any word to be accessed from the disk with only one disk access operation.…”
Section: Introductionmentioning
confidence: 99%
“…After further investigation we developed a modified algorithm [16] requiring O(m 3) time. With this algorithm we were able to find MPHF's for sets of over a thousand words.…”
Section: Practical Minimal Perfect Nash Functions For Large Databasesmentioning
confidence: 99%
“…After careful analysis of Sager's algorithm [32] and our enhanced version [16], Heath made three crucial observations that serve as the foundation of our new algorithms:…”
Section: Key Concepts Of New Algorithmsmentioning
confidence: 99%