Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP) 2014
DOI: 10.3115/v1/d14-1055
|View full text |Cite
|
Sign up to set email alerts
|

Positive Unlabeled Learning for Deceptive Reviews Detection

Abstract: Deceptive reviews detection has attracted significant attention from both business and research communities. However, due to the difficulty of human labeling needed for supervised learning, the problem remains to be highly challenging. This paper proposed a novel angle to the problem by modeling PU (positive unlabeled) learning. A semi-supervised model, called mixing population and individual property PU learning (MPIPUL), is proposed. Firstly, some reliable negative examples are identified from the unlabeled … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
39
1

Year Published

2015
2015
2022
2022

Publication Types

Select...
4
2
2

Relationship

0
8

Authors

Journals

citations
Cited by 70 publications
(44 citation statements)
references
References 26 publications
0
39
1
Order By: Relevance
“…In Table 4 we can observe the indirect comparison of our results with those of (Banerjee and Chua, 2014) and (Ren et al, 2014) obtained with a 10 fold cross validation experiment, and then, with a 5 fold cross validation in order to make a fair comparison with the results of (Ott et al, 2011) and (Feng and Hirst, 2013). Note that the results are expressed in terms of the accuracy as those were published by the authors; the results correspond only to positive reviews of the Opinion Spam corpus because the authors experimented in that corpus alone.…”
Section: Comparison Of Resultsmentioning
confidence: 91%
See 2 more Smart Citations
“…In Table 4 we can observe the indirect comparison of our results with those of (Banerjee and Chua, 2014) and (Ren et al, 2014) obtained with a 10 fold cross validation experiment, and then, with a 5 fold cross validation in order to make a fair comparison with the results of (Ott et al, 2011) and (Feng and Hirst, 2013). Note that the results are expressed in terms of the accuracy as those were published by the authors; the results correspond only to positive reviews of the Opinion Spam corpus because the authors experimented in that corpus alone.…”
Section: Comparison Of Resultsmentioning
confidence: 91%
“…The accuracy of the semi-supervised model is slightly lower (86.69%) than that of our approach (89%), although good enough. The authors concluded that the good performance of the semi-supervised model is due the topic information captured by the model combined with the examples and their similarity (Ren et al, 2014). Then, they could obtain an accurate SVM classifier.…”
Section: Comparison Of Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…Prior to that, researchers such as Jindal and Liu [ 13 ] implemented a machine learning approach by adopting the logistic regression to detect opinion spams, so did Li, Ott [ 14 ], Lin, Zhu [ 15 ] and Ren and Ji [ 16 ] who likewise, used supervised machine learning to detect opinion spams. However, earlier studies by Ren, Ji [ 17 ] and Li, Chen [ 18 ] implemented a hybrid approach that combined supervised and semi-supervised machine learning to detect opinion spams. Despite this being so, there were several limitations to the machine learning approach, for example, using too many features, providing less accurate outcomes, poor flexibility and high consumption of computational time.…”
Section: Introductionmentioning
confidence: 99%
“…Contrary to the positive-only case, Positive and Unlabeled (PU) learning [8,17,18,27,29], leverages some non-labeled data to further improve the performance of the model in classification tasks. One of the classical methods is to identify possible negatives in this unlabeled set.…”
mentioning
confidence: 99%