Today, Artificial Intelligence is achieving prodigious real-time performance, thanks to growing computational data and power capacities. However, there is little knowledge about what system results convey; thus, they are at risk of being susceptible to bias, and with the roots of Artificial Intelligence (“AI”) in almost every territory, even a minuscule bias can result in excessive damage. Efforts towards making AI interpretable have been made to address fairness, accountability, and transparency concerns. This paper proposes two unique methods to understand the system’s decisions aided by visualizing the results. For this study, interpretability has been implemented on Natural Language Processing-based sentiment analysis using data from various social media sites like Twitter, Facebook, and Reddit. With Valence Aware Dictionary for Sentiment Reasoning (“VADER”), heatmaps are generated, which account for visual justification of the result, increasing comprehensibility. Furthermore, Locally Interpretable Model-Agnostic Explanations (“LIME”) have been used to provide in-depth insight into the predictions. It has been found experimentally that the proposed system can surpass several contemporary systems designed to attempt interpretability.
Social platforms such as Twitter and Facebook have now become only media to express their thoughts, and due to lack of censorship, it often embellishes themselves as an abode for hate towards minorities. People of color, Asian people, Muslims, women, transgenders, and LGBTQ+ communities are often the target of such online hate and aggression. Though several companies have incorporated considerable algorithms on their platforms, nevertheless due to being rather hard to often detect such comments still make it to the platforms, creating a negative space towards targeted people. This research involves the study and comparison of different hate and aggression detection algorithms with intent on two languages, i.e. English and German including machine learning models (linear SVC, logistic regression, multinomial naive Bayes and random forests) with their variations with feature engineering and bag of words and deep learning (CNN-GRU static, TCN static, Seq2Seq) with their variations vis-à-vis Word2Vec embedding. CNN+GRU static + Word2Vec embedding has outperformed all the other techniques with an accuracy of 68.29%.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.