Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing 2018
DOI: 10.18653/v1/d18-1383
|View full text |Cite
|
Sign up to set email alerts
|

Adaptive Semi-supervised Learning for Cross-domain Sentiment Classification

Abstract: We consider the cross-domain sentiment classification problem, where a sentiment classifier is to be learned from a source domain and to be generalized to a target domain. Our approach explicitly minimizes the distance between the source and the target instances in an embedded feature space. With the difference between source and target minimized, we then exploit additional information from the target domain by consolidating the idea of semi-supervised learning, for which, we jointly employ two regularizations… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
50
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 56 publications
(54 citation statements)
references
References 17 publications
0
50
0
Order By: Relevance
“…Beauty, Book, Music: These data sets each consist of 6000 beauty product, book and music reviews with a balanced distribution of three class labels. This data set was introduced by (He et al, 2018). Data is obtained by sampling from a larger data set of product reviews obtained from Amazon (McAuley et al, 2015).…”
Section: Resultsmentioning
confidence: 99%
“…Beauty, Book, Music: These data sets each consist of 6000 beauty product, book and music reviews with a balanced distribution of three class labels. This data set was introduced by (He et al, 2018). Data is obtained by sampling from a larger data set of product reviews obtained from Amazon (McAuley et al, 2015).…”
Section: Resultsmentioning
confidence: 99%
“…In our implementation, the feature encoder G consists of three parts including a 300-dimensional word embedding layer using GloVe (Pennington et al, 2014), a one-layer CNN with ReLU activation function adopted in (Yu and Jiang, 2016;He et al, 2018) and a max-over-time pooling through which final sentence representation is obtained. Specifically, the convolution filter and the window size of this one-layer CNN are 300 and 3 separately.…”
Section: Training Details and Hyper-parametersmentioning
confidence: 99%
“…During training period, λ 1 , λ 2 , λ 3 , λ 4 , and n are set to 5.0, 0.1, 0.1, 1.5, 2. Similar to (He et al, 2018), we parametrize λ 4 as a dynamic weight exp[−5(1 − t max−epochs ) 2 ]λ 4 . This is to minimize the effort of the regularizer as the predictor is not good at the beginning of training.…”
Section: Training Details and Hyper-parametersmentioning
confidence: 99%
“…Existing relevant studies could be attributed into two categories: two-stage approaches (Blitzer, Dredze, and Pereira 2007;Glorot, Bordes, and Bengio 2011;Ziser and Reichart 2018; and end-to-end models (Ganin et al 2016;He et al 2018;Qu et al 2019). The two-stage approaches typically construct unsupervised feature extractors or manually select pivot features across domains in the first stage.…”
Section: Introductionmentioning
confidence: 99%
“…That is, their sentiment classifiers and feature extractors overlook the sentiment polarity lying in the review text of target domains. One relevant study (He et al 2018) in this regard obtains pseudo-labels in target domains by a self-ensemble bootstrapping technique to train its sentiment classifier and feature extractor. However, the pseudo-labels are asynchronously generated by the earlier version of the sentiment classifier, which is a weaker classifier compared with its current version, and possibly limits the effectiveness of training.…”
Section: Introductionmentioning
confidence: 99%