Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics 2020
DOI: 10.18653/v1/2020.acl-main.96
|View full text |Cite
|
Sign up to set email alerts
|

Code-Switching Patterns Can Be an Effective Route to Improve Performance of Downstream NLP Applications: A Case Study of Humour, Sarcasm and Hate Speech Detection

Abstract: In this paper we demonstrate how codeswitching patterns can be utilised to improve various downstream NLP applications. In particular, we encode different switching features to improve humour, sarcasm and hate speech detection tasks. We believe that this simple linguistic observation can also be potentially helpful in improving other similar NLP applications.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
1
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 13 publications
(6 citation statements)
references
References 9 publications
0
5
0
Order By: Relevance
“…Recent literature has made significant efforts to understand syntactic structure and semantics from code-mixed texts [3,4,5]. Similar attempts have been made for pragmatic tasks -humour, sarcasm and hate detection in the code-mixed regime [6,7].…”
Section: Introductionmentioning
confidence: 94%
See 1 more Smart Citation
“…Recent literature has made significant efforts to understand syntactic structure and semantics from code-mixed texts [3,4,5]. Similar attempts have been made for pragmatic tasks -humour, sarcasm and hate detection in the code-mixed regime [6,7].…”
Section: Introductionmentioning
confidence: 94%
“…: We perform a human evaluation study to evaluate the code-mixed texts generated by PARADOX and the vanilla Transformer. We randomly sample 24 examples from each of these models and ask 30 human evaluators 7 to rate these examples based on Semantic coherence and Linguistic quality. Semantic coherence measures the meaningfulness of the code-mixed texts, whereas, with linguistic quality, we measure their structural validity.…”
Section: Comparative Analysismentioning
confidence: 99%
“…This can be combated by collecting data from platforms with a primary Urdu-speaking population, or by considering the conversational data between people who communicate in Urdu with each other, as may be the case of native Urdu-speaking families. Alternatively, a modern Deep Neural Model, using datasets based on theme, was used by Bansal et al (2020) to improve upon the accuracy achieved by other models. The research presented a set of nine features, which could be further improved by the use of switching features in the final layer of the Deep Network.…”
Section: Code-switchingmentioning
confidence: 99%
“…We recommend to specifically investigate how to integrate these characteristics into future models. For instance, Alorainy et al [9] extract features specifically to identify othering language, Bansal et al [26] and recent publications in ACL workshops [19] focus on humor and sarcasm. Discriminatory features.…”
Section: Lack Of Ocl-dependent Featuresmentioning
confidence: 99%