“…This is mainly due to availability of a large set of samples of ironic texts, which are easy to be collected relying on the behavior of Twitter users, who often explicitly mark their ironic messages by using hashtags such as '#irony' or '#sarcasm'. The pretty good reliability of the user-generated hashtags as golden labels for irony has been experimentally confirmed by [Kunneman et al 2015]. Moreover, it seems that, due to the interaction model underlying the microblogging platform, irony expressed here could be somehow easier to analyze.…”
Section: Related Workmentioning
confidence: 54%
“…Furthermore, there are some efforts in other so-cial media such as customer reviews from Amazon 1 [Filatova 2012;Buschmeier et al 2014]; comments from the online debate sites such as 4forums.com 2 [Abbott et al 2011;Lukin and Walker 2013] and, recently, Reddit 3 . The majority of the research in irony detection has been addressed in English, although there is some research in other languages, such as: Dutch [Kunneman et al 2015], Italian [Bosco et al 2013], Czech [Ptáček et al 2014], French [Karoui et al 2015], Portuguese [Carvalho et al 2009] and Chinese [Tang and Chen 2014]. A shared task for English on sentiment analysis of figurative language in Twitter has been organized at SemEval-2015 for the first time [Ghosh et al 2015], and a pilot shared task for Italian on irony detection has been proposed in Sentipolc-2014 within the periodic evaluation campaign EVALITA [Basile et al 2014;Attardi et al 2015].…”
Irony has been proven to be pervasive in social media, posing a challenge to sentiment analysis systems. It is a creative linguistic phenomenon where affect-related aspects play a key role. In this work, we address the problem of detecting irony in tweets, casting it as a classification problem. We propose a novel model which explores the use of affective features based on a wide range of lexical resources available for English, reflecting different facets of affect. Classification experiments over different corpora show that affective information helps in distinguishing among ironic and non-ironic tweets. Our model outperforms the stateof-the-art in almost all cases.
“…This is mainly due to availability of a large set of samples of ironic texts, which are easy to be collected relying on the behavior of Twitter users, who often explicitly mark their ironic messages by using hashtags such as '#irony' or '#sarcasm'. The pretty good reliability of the user-generated hashtags as golden labels for irony has been experimentally confirmed by [Kunneman et al 2015]. Moreover, it seems that, due to the interaction model underlying the microblogging platform, irony expressed here could be somehow easier to analyze.…”
Section: Related Workmentioning
confidence: 54%
“…Furthermore, there are some efforts in other so-cial media such as customer reviews from Amazon 1 [Filatova 2012;Buschmeier et al 2014]; comments from the online debate sites such as 4forums.com 2 [Abbott et al 2011;Lukin and Walker 2013] and, recently, Reddit 3 . The majority of the research in irony detection has been addressed in English, although there is some research in other languages, such as: Dutch [Kunneman et al 2015], Italian [Bosco et al 2013], Czech [Ptáček et al 2014], French [Karoui et al 2015], Portuguese [Carvalho et al 2009] and Chinese [Tang and Chen 2014]. A shared task for English on sentiment analysis of figurative language in Twitter has been organized at SemEval-2015 for the first time [Ghosh et al 2015], and a pilot shared task for Italian on irony detection has been proposed in Sentipolc-2014 within the periodic evaluation campaign EVALITA [Basile et al 2014;Attardi et al 2015].…”
Irony has been proven to be pervasive in social media, posing a challenge to sentiment analysis systems. It is a creative linguistic phenomenon where affect-related aspects play a key role. In this work, we address the problem of detecting irony in tweets, casting it as a classification problem. We propose a novel model which explores the use of affective features based on a wide range of lexical resources available for English, reflecting different facets of affect. Classification experiments over different corpora show that affective information helps in distinguishing among ironic and non-ironic tweets. Our model outperforms the stateof-the-art in almost all cases.
“…Also, it was observed that harsh tweets contained more secondperson pronouns (1.18% of all tokens) than non-harsh tweets (0.59%). As shown by among others Kunneman et al (2015) and Riloff et al (2013), ironic utterances are likely to contain markers such as interjections (e.g. 'yeah right') or intensifiers and diminishers to express hyperbole and understatements.…”
Section: Corpus Analysismentioning
confidence: 91%
“…They distinguished tweets with the #irony hashtag from tweets with the hashtags #education, #humor or #politics and found that the best scores were obtained by distinguishing #irony from #humor tweets. Kunneman et al (2015) pioneered irony detection in Dutch tweets using word n-gram features and a Balanced Winnow classifier. They trained a classifier by contrasting the hashtag-labelled irony tweets (i.e.…”
Section: Computational Approaches To Ironymentioning
confidence: 99%
“…#irony, #sarcasm, #not) assigned by the author of the text to label instances in an irony corpus, but this has shown to introduce noise into the labelled training data (Kunneman et al 2015). For the current study, we also collected ironic tweets using the above hashtags, but supplemented them with manual annotations using a finegrained annotation scheme (Van Hee et al 2016b).…”
To push the state of the art in text mining applications, research in natural language processing has increasingly been investigating automatic irony detection, but manually annotated irony corpora are scarce. We present the construction of a manually annotated irony corpus based on a fine-grained annotation scheme that allows for identification of different types of irony. We conduct a series of binary classification experiments for automatic irony recognition using a support vector machine (SVM) that exploits a varied feature set and compare this method to a deep learning approach that is based on an LSTM network and (pre-trained) word embeddings. Evaluation on a held-out corpus shows that the SVM model outperforms the neural network approach and benefits from combining lexical, semantic and syntactic information sources. A qualitative analysis of the classification output reveals that the classifier performance may be further enhanced by integrating implicit sentiment information and context-and user-based features.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.