1993
DOI: 10.1093/llc/8.1.20
|View full text |Cite
|
Sign up to set email alerts
|

Characteristics of Sentence Length in Running Text

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2010
2010
2021
2021

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 9 publications
(6 citation statements)
references
References 0 publications
0
6
0
Order By: Relevance
“…Shot lengths were analyzed using partial autocorrelation and power analyses, which allowed us to look for local patterns (shot-to-shot relations) and global patterns (whole-film editing profiles), respectively. Schils and de Haan (1993) performed a similar local analysis on sentence lengths in texts, and Salt (2006, p. 396) provided some piecemeal, local analyses of a number of films. In addition, Richards, Wilson, and Sommer (1994, Experiment 4) analyzed portions of four films in a manner related to our global analysis.…”
Section: Film Choice Shot Parsing and Analysismentioning
confidence: 99%
“…Shot lengths were analyzed using partial autocorrelation and power analyses, which allowed us to look for local patterns (shot-to-shot relations) and global patterns (whole-film editing profiles), respectively. Schils and de Haan (1993) performed a similar local analysis on sentence lengths in texts, and Salt (2006, p. 396) provided some piecemeal, local analyses of a number of films. In addition, Richards, Wilson, and Sommer (1994, Experiment 4) analyzed portions of four films in a manner related to our global analysis.…”
Section: Film Choice Shot Parsing and Analysismentioning
confidence: 99%
“…Three main challenges arose. (1) Proteins range from about 30 to 33,000 residues, a much larger range than for the average English sentence extending over 15–30 words [44], and even more extreme than notable literary exceptions such as James Joyce’s Ulysses (1922) with almost 4000 words in a sentence. Longer proteins require more GPU memory and the underlying models (so-called LSTMs: Long Short-Term Memory networks [45]) have only a limited capability to remember long-range dependencies.…”
Section: Introductionmentioning
confidence: 99%
“…In this study, we introduced basic algorithms and reviewed the recent literature concerning representation learning applications in sequence analysis. Heinzinger, et al highlighted three difficulties in biological sequence modeling with NLP [68] as follows: (i) proteins range from approximately 30 to 33,000 residues, which is markedly longer than the average English sentence, which consists of 15 to 30 words [106] ; (ii) proteins use only 20 amino acids in most cases; if we consider one amino acid as a word, the word repertoire is 1/100,000 of English language, and if we consider 3-mer as a word, the word repertoire is 1/10 to 1/100 of English language; (iii) UniProt [90] is 10 times larger than the size of Wikipedia in terms of data repository size, and extracting information from a very large biological database may require the use of a commensurate model. Embedding of biological sequences using NLP overcomes these difficulties and outperforms existing methods in several tasks, such as function, structure, localization, and disorder prediction ( Table 1 ).…”
Section: Discussionmentioning
confidence: 99%