Abstract:Social media users, including organizations, often struggle to acquire the maximum number of responses from other users, but predicting the responses that a post will receive before publication is highly desirable. Previous studies have analyzed why a given tweet may become more popular than others, and have used a variety of models trained to predict the response that a given tweet will receive. The present research addresses the prediction of response measures available on Twitter, including likes, replies a… Show more
“…The only case in which DL methods provided low performance was that of a pre-trained Roberta large model without any training on the data set in question. By comparing our results with those for similar methods applied to other Twitter text analysis domains, such as sentiment analysis [24] or the prediction of user response to Twitter posts [24] we find that the order of performance of the used methods agrees with those from other research: recent DL methods fine-tuned in a supervised manner on problem-specific data provide the highest prediction quality.…”
Section: A Overall Methods Performancesupporting
confidence: 78%
“…Twitter language is also known to have both oral and written language characteristics as debated by Wikström in [22], and Garber in [23], and the language of tweets addressed to USNTA demonstrates numerous elements of military specifics and jargon, for example, the use of acronyms such as "BZ" (Bravo Zulu), "SAPR survey" (Sexual Assault Prevention and Response survey), or "TADMUS" (tactical decision-making under stress). In this research, we have used tweets published in a previous Twitter-related study regarding sentiment analysis by Fiok et al [24].…”
Efficient leadership plays an important role in organizations, with the military being one of the more obvious examples of this statement. In this context, it is not surprising that ensuring a culture of excellence is at the heart of Navy leadership. However, it is not easy to maintain or increase the quality of leadership among staff, as such efforts require constant training and practice. To address this need for continuous monitoring and improvement in human leadership expressed in everyday communication, we demonstrate the feasibility of automatically detecting and classifying military leadership messages. We achieve this goal by 1) curating a data set of short text messages that are written in the military-specific language, have some characteristics of spoken language, and are human-annotated with labels referring to selected leadership roles and 2) demonstrating the performance of selected automation methods that allow classes to be predicted for each analyzed message. This study shows that recent deep learning methods provide reasonable performance, even when limited data is provided. Future efforts should focus on creating an automated self-assessment tool that would enable continuous monitoring and training of leadership skills required in the Navy domain.
“…The only case in which DL methods provided low performance was that of a pre-trained Roberta large model without any training on the data set in question. By comparing our results with those for similar methods applied to other Twitter text analysis domains, such as sentiment analysis [24] or the prediction of user response to Twitter posts [24] we find that the order of performance of the used methods agrees with those from other research: recent DL methods fine-tuned in a supervised manner on problem-specific data provide the highest prediction quality.…”
Section: A Overall Methods Performancesupporting
confidence: 78%
“…Twitter language is also known to have both oral and written language characteristics as debated by Wikström in [22], and Garber in [23], and the language of tweets addressed to USNTA demonstrates numerous elements of military specifics and jargon, for example, the use of acronyms such as "BZ" (Bravo Zulu), "SAPR survey" (Sexual Assault Prevention and Response survey), or "TADMUS" (tactical decision-making under stress). In this research, we have used tweets published in a previous Twitter-related study regarding sentiment analysis by Fiok et al [24].…”
Efficient leadership plays an important role in organizations, with the military being one of the more obvious examples of this statement. In this context, it is not surprising that ensuring a culture of excellence is at the heart of Navy leadership. However, it is not easy to maintain or increase the quality of leadership among staff, as such efforts require constant training and practice. To address this need for continuous monitoring and improvement in human leadership expressed in everyday communication, we demonstrate the feasibility of automatically detecting and classifying military leadership messages. We achieve this goal by 1) curating a data set of short text messages that are written in the military-specific language, have some characteristics of spoken language, and are human-annotated with labels referring to selected leadership roles and 2) demonstrating the performance of selected automation methods that allow classes to be predicted for each analyzed message. This study shows that recent deep learning methods provide reasonable performance, even when limited data is provided. Future efforts should focus on creating an automated self-assessment tool that would enable continuous monitoring and training of leadership skills required in the Navy domain.
“…Beyond single tweet classification, tweet popularitylikes, retweets, number of replies, etc.-has been studied using account information (Matsumoto et al 2019;Fiok et al 2020) and linguistic information (Wang, Chen, and Kan 2012). Hate and counter hate have also been studied beyond single tweets.…”
Recent studies in the hate and counter hate domain have provided the grounds for investigating how to detect this pervasive content in social media. These studies mostly work with synthetic replies to hateful content written by annotators on demand rather than replies written by real users. We argue that working with naturally occurring replies to hateful content is key to study the problem. Building on this motivation, we create a corpus of 5,652 hateful tweets and replies. We analyze their fine-grained relationships by indicating whether the reply (a) is hate or counter hate speech, (b) provides a justification, (c) attacks the author of the tweet, and (d) adds additional hate. We also present linguistic insights into the language people use depending on these fine-grained relationships. Experimental results show improvements (a) taking into account the hateful tweet in addition to the reply and (b) pretraining with related tasks.
“…Predicting popularity of messages is a straightforward task from the perspective of machine learning and has been framed both as a regression (Lampos et al, 2014) and classification problem (Hong et al, 2011;Jenders et al, 2013;Subramanian et al, 2018;Fiok et al, 2020), while work on information cascades (Zhao et al, 2015;Li et al, 2017;Zhou et al, 2021) focuses on modeling the entire lifetime of a post as a point process.…”
Twitter has slowly but surely established itself as a forum for disseminating, analysing and promoting NLP research. The trend of researchers promoting work not yet peerreviewed (preprints) by posting concise summaries presented itself as an opportunity to collect and combine multiple modalities of data. In scope of this paper, we (1) construct a dataset of Twitter threads in which researchers promote NLP preprints and (2) evaluate whether it is possible to predict the popularity of a thread based on the content of the Twitter thread, paper content and user metadata. We experimentally show that it is possible to predict popularity of threads promoting research based on their content, and that predictive performance depends on modelling textual input, indicating that the dataset could present value for related areas of NLP research such as citation recommendation and abstractive summarization.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.