2020
DOI: 10.3390/sci2040078
|View full text |Cite
|
Sign up to set email alerts
|

Statistics and Machine Learning Experiments in English and Romanian Poetry

Abstract: This paper presents a quantitative approach to poetry, based on the use of several statistical measures (entropy, informational energy, N-gram, etc.) applied to a few characteristic English writings. We found that English language changes its entropy as time passes, and that entropy depends on the language used and on the author. In order to compare two similar texts, we were able to introduce a statistical method to asses the information entropy between two texts. We also introduced a method of computing the … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(1 citation statement)
references
References 32 publications
0
1
0
Order By: Relevance
“…Despite the existence of other models, for example described by Chomsky (1956), many studies of natural languages, and in particular English, use the approximation of the text by the stationary Marcov process. For example, in the papers of Calin (2020), Hahn andSivley (2011), Yadav et al (2010), Guerrero (2009) the Marcov process is used to simulate a natural language text. Since the accuracy of approximating a natural language text using the Marcov model decreases significantly with an increase in the order of n-gram, related studies mainly investigate the entropy of short-length texts.…”
Section: Related Workmentioning
confidence: 99%
“…Despite the existence of other models, for example described by Chomsky (1956), many studies of natural languages, and in particular English, use the approximation of the text by the stationary Marcov process. For example, in the papers of Calin (2020), Hahn andSivley (2011), Yadav et al (2010), Guerrero (2009) the Marcov process is used to simulate a natural language text. Since the accuracy of approximating a natural language text using the Marcov model decreases significantly with an increase in the order of n-gram, related studies mainly investigate the entropy of short-length texts.…”
Section: Related Workmentioning
confidence: 99%