2017
DOI: 10.3390/e19070299
|View full text |Cite
|
Sign up to set email alerts
|

Critical Behavior in Physics and Probabilistic Formal Languages

Abstract: Abstract:We show that the mutual information between two symbols, as a function of the number of symbols between the two, decays exponentially in any probabilistic regular grammar, but can decay like a power law for a context-free grammar. This result about formal languages is closely related to a well-known result in classical statistical mechanics that there are no phase transitions in dimensions fewer than two. It is also related to the emergence of power law correlations in turbulence and cosmological infl… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

2
111
0
1

Year Published

2018
2018
2021
2021

Publication Types

Select...
4
1
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 74 publications
(121 citation statements)
references
References 54 publications
2
111
0
1
Order By: Relevance
“…The TRW of the language areas therefore appears to be in the 5-7 word range. As alluded to in the Introduction, this relatively local linguistic processing is likely driven by the statistical properties of natural language, where most semantic/syntactic dependencies are local (e.g., Futrell et al, 2015), and PMI falls off quite sharply as a function of inter-word distance (e.g., Lin & Tegmark, 2017). We can further speculate that linguistic chunks of this size are sufficient to express clause-level meanings, where clauses describe events -salient and meaningful semantic units in our experience with the world (e.g., Zacks & Tversky, 2001).…”
Section: The Temporal Receptive Window Of the Language Areasmentioning
confidence: 99%
See 1 more Smart Citation
“…The TRW of the language areas therefore appears to be in the 5-7 word range. As alluded to in the Introduction, this relatively local linguistic processing is likely driven by the statistical properties of natural language, where most semantic/syntactic dependencies are local (e.g., Futrell et al, 2015), and PMI falls off quite sharply as a function of inter-word distance (e.g., Lin & Tegmark, 2017). We can further speculate that linguistic chunks of this size are sufficient to express clause-level meanings, where clauses describe events -salient and meaningful semantic units in our experience with the world (e.g., Zacks & Tversky, 2001).…”
Section: The Temporal Receptive Window Of the Language Areasmentioning
confidence: 99%
“…The degree of local combinability can be formally estimated using tools from information theory (Shannon & Weaver, 1963). Naturalistic linguistic input is characterized by relatively high pointwise mutual information (PMI) among words within a local linguistic context, and it falls off for word pairs spanning longer distances (e.g., Li, 1990;Lin & Tegmark, 2017;Futrell, Qian, Gibson, Fedorenko, & Blank, 2019). Our local-word-swap manipulation maintained approximately the same level of local mutual information as that observed in typical linguistic input.…”
mentioning
confidence: 90%
“…They show that an irreducible and aperiodic Markov process, with non-degenerate eigenvalues, cannot produce critical behaviour because I decays exponentially. This phenomenon is seen in a number of cases, including hidden and semi-Markov models 1,25 . In the literature, such behaviour is superficially dealt with by increasing the state space to include symbols from the past, which does not address the main issue 25 with Markov models; lack of memory.…”
Section: /15mentioning
confidence: 93%
“…This phenomenon is seen in a number of cases, including hidden and semi-Markov models 1,25 . In the literature, such behaviour is superficially dealt with by increasing the state space to include symbols from the past, which does not address the main issue 25 with Markov models; lack of memory. This analysis shows that GeoLife dataset consists of considerably higher number of long-range correlations, compared to the PrivaMov dataset and the NMDC dataset.…”
Section: /15mentioning
confidence: 93%
“…Their claim that the LSTM model is capable of capturing longrange dependencies is thus only supported by such qualitative evidence, without giving a deep insight in the characteristics of the generated documents. Lin and Tegmark [16] compared natural language texts with those generated by Markov models and LSTMs, exploiting metrics coming from information theory. Their analysis shows that LSTMs are capable of capturing correlations that Markov models instead fail to represent, yet the range of correlations they consider is still quite limited (up to 1,000 characters).…”
Section: Introductionmentioning
confidence: 99%