2018
DOI: 10.1007/978-981-10-8438-6_23
|View full text |Cite
|
Sign up to set email alerts
|

Detecting Computer-Generated Text Using Fluency and Noise Features

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
1

Relationship

2
3

Authors

Journals

citations
Cited by 5 publications
(4 citation statements)
references
References 7 publications
0
4
0
Order By: Relevance
“…In our previous work [9], we extracted word density features using an N -gram language model on both internally limited corpus and huge external corpus. Futhermore, we found that the human-generated text frequently contains particular words such as spoken words (e.g., wanna, gonna) or misspelling words (comin, goin, etc.)…”
Section: B Sentence Levelmentioning
confidence: 99%
See 1 more Smart Citation
“…In our previous work [9], we extracted word density features using an N -gram language model on both internally limited corpus and huge external corpus. Futhermore, we found that the human-generated text frequently contains particular words such as spoken words (e.g., wanna, gonna) or misspelling words (comin, goin, etc.)…”
Section: B Sentence Levelmentioning
confidence: 99%
“…Our previous method extracted two features from informal text at the sentence level: a density feature using an N -gram language model and a noise feature to be matched unexpected words (misspelling words, translated error words, etc.) with original forms of words included in the standard lexica [9]. The drawback of this method is that, however, these unexpected words are easily recognized and corrected by advanced assistant tools in formal text (e.g.…”
Section: Introductionmentioning
confidence: 99%
“…The first approach extracted distinguishable features from parsing trees [3,6], but such trees are only parsed from an individual sentence. To overcome this problem, other methods [1,2,8] based on N -gram language model extract such features from nearby words in both inside and outside a sentence. The limitation of this model is that meaningful features are only given from few nearby words, common in three.…”
Section: Machinetranslated Paragraph Pmmentioning
confidence: 99%
“…There are two other reasonable combinations also aim to diminish the restriction of N -gram model. The first combination [8] extracted the specific noise words often used by a human, that is misspelled and reduction words, or by a machine, namely untranslated words. This combination, however, is only efficient in online social network in which contains a substantial number of such noises.…”
Section: N -Gram Modelmentioning
confidence: 99%