Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing 2015
DOI: 10.18653/v1/d15-1282
|View full text |Cite
|
Sign up to set email alerts
|

Social Media Text Classification under Negative Covariate Shift

Abstract: In a typical social media content analysis task, the user is interested in analyzing posts of a particular topic. Identifying such posts is often formulated as a classification problem. However, this problem is challenging. One key issue is covariate shift. That is, the training data is not fully representative of the test data. We observed that the covariate shift mainly occurs in the negative data because topics discussed in social media are highly diverse and numerous, but the user-labeled negative training… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
35
0

Year Published

2016
2016
2023
2023

Publication Types

Select...
5
1
1

Relationship

1
6

Authors

Journals

citations
Cited by 32 publications
(42 citation statements)
references
References 39 publications
0
35
0
Order By: Relevance
“…Second, it only focuses on improving accuracy on the classes with seed examples. Our work in [14] dealt with the problem using an entirely different approach adopted from [13]. However, these works did not propose or deal with cumulative learning, which is important for an intelligent system as it allows the system to learn more and more and become more and more knowledgeable.…”
Section: Open World Classificationmentioning
confidence: 99%
See 3 more Smart Citations
“…Second, it only focuses on improving accuracy on the classes with seed examples. Our work in [14] dealt with the problem using an entirely different approach adopted from [13]. However, these works did not propose or deal with cumulative learning, which is important for an intelligent system as it allows the system to learn more and more and become more and more knowledgeable.…”
Section: Open World Classificationmentioning
confidence: 99%
“…At time + 1, the new dataset +1 of class +1 arrives, and the classification model needs to be updated or extended to produce a new classification model +1 . We note that each ℎ in or +1 is a 1-vs-rest SVM classifier built using the CBS learning method in [13] for class treating as the positive class. We will discuss CBS learning in the next section.…”
Section: Training a Cumulative Classification Modelmentioning
confidence: 99%
See 2 more Smart Citations
“…The authors in [44] reported that a linear SVM achieves the best results consistently to SVM with different kernels including SVM-Poly. The authors in [7,16,26] also reported the linear SVM efficiency for binary text classification. Unfortunately, manual hyperparameter selection still remains one of the practical application issues, while recent literature still does not provide any heuristic rules or rules of thumb for this task in [41].…”
Section: Introductionmentioning
confidence: 99%