Proceedings of the 10th International Workshop on Semantic Evaluation (SemEval-2016) 2016
DOI: 10.18653/v1/s16-1067
|View full text |Cite
|
Sign up to set email alerts
|

DeepStance at SemEval-2016 Task 6: Detecting Stance in Tweets Using Character and Word-Level CNNs

Abstract: This paper describes our approach for the Detecting Stance in Tweets task (SemEval-2016 Task 6). We utilized recent advances in short text categorization using deep learning to create word-level and character-level models. The choice between word-level and characterlevel models in each particular case was informed through validation performance. Our final system is a combination of classifiers using word-level or character-level models. We also employed novel data augmentation techniques to expand and diversif… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
29
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 40 publications
(29 citation statements)
references
References 10 publications
0
29
0
Order By: Relevance
“…Cosine distance between embeddings of reference source tweets and those of unlabeled candidate tweets is used as a measurement of semantic similarity. Cosine similarity between vector representation of two sentences is a common metric for measuring semantic similarity [20]. Two semantically equivalent embeddings have a cosine similarity of 1, and two vectors with no relation have that of 0.…”
Section: A Overview Of the Proposed Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…Cosine distance between embeddings of reference source tweets and those of unlabeled candidate tweets is used as a measurement of semantic similarity. Cosine similarity between vector representation of two sentences is a common metric for measuring semantic similarity [20]. Two semantically equivalent embeddings have a cosine similarity of 1, and two vectors with no relation have that of 0.…”
Section: A Overview Of the Proposed Methodsmentioning
confidence: 99%
“…In particular, character-level CNNs trained on augmented data achieves the best performance. Recent research [19,20] applies this method to tweets, and shows that data augmentation can bring performance gains in deep learning tasks on noisy and short social media texts. Vosoughi et al [19] augment domainindependent English tweets for training an encoder-decoder embedding model built with character-level CNN and long short-term memory (LSTM).…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…MITRE [14] provided the best deep learning solution in the contest, initializing weights from a 256-dimensional word embeddings learned using the word2vec skip-gram algorithm [6], followed by a second layer with 128 LSTM units. Among others, pkudblab [12] and DeepStance [11] use deep CNN models. Augenstein et al [1] employ a bidirectional attention model.…”
Section: Related Workmentioning
confidence: 99%
“…Task A External resources: Bag-of-Words and word vectors. DeepStance [19] Overall approach: A set of naive bayes classifiers using deep learning.…”
Section: Our Approachmentioning
confidence: 99%