2021
DOI: 10.3390/e23091148
|View full text |Cite
|
Sign up to set email alerts
|

A Refutation of Finite-State Language Models through Zipf’s Law for Factual Knowledge

Abstract: We present a hypothetical argument against finite-state processes in statistical language modeling that is based on semantics rather than syntax. In this theoretical model, we suppose that the semantic properties of texts in a natural language could be approximately captured by a recently introduced concept of a perigraphic process. Perigraphic processes are a class of stochastic processes that satisfy a Zipf-law accumulation of a subset of factual knowledge, which is time-independent, compressed, and effectiv… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2

Citation Types

0
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
2
1

Relationship

1
2

Authors

Journals

citations
Cited by 3 publications
(4 citation statements)
references
References 104 publications
0
4
0
Order By: Relevance
“…Seen in this light, Theorems 3 and 4 make a link between Hilberg's hypothesis about a power-law growth of mutual information for natural language [22,7] and Herdan-Heaps' law about the power-law growth of the number of distinct words in a text [19,17], which is a corollary of Zipf's law. For more details, see [10,11,12].…”
Section: Proof Of Theoremmentioning
confidence: 99%
See 3 more Smart Citations
“…Seen in this light, Theorems 3 and 4 make a link between Hilberg's hypothesis about a power-law growth of mutual information for natural language [22,7] and Herdan-Heaps' law about the power-law growth of the number of distinct words in a text [19,17], which is a corollary of Zipf's law. For more details, see [10,11,12].…”
Section: Proof Of Theoremmentioning
confidence: 99%
“…In particular, we may estimate the Markov order of empirical data using a consistent estimator, such as one introduced by Merhav, Gutman, and Ziv [25], and we may count the number of distinct substrings of the length equal to the Markov order estimate. In such case, we may observe that natural language contains many more such substrings than memoryless sources, see [11,12].…”
Section: Proof Of Theoremmentioning
confidence: 99%
See 2 more Smart Citations