2020
DOI: 10.1007/s40264-020-00912-9
|View full text |Cite
|
Sign up to set email alerts
|

Adverse Events in Twitter-Development of a Benchmark Reference Dataset: Results from IMI WEB-RADR

Abstract: Introduction and Objective Social media has been suggested as a source for safety information, supplementing existing safety surveillance data sources. This article summarises the activities undertaken, and the associated challenges, to create a benchmark reference dataset that can be used to evaluate the performance of automated methods and systems for adverse event recognition. Methods A retrospective analysis of public English-language Twitter posts (Tweets) was performed. We sampled 57,473 Tweets out of 5,… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
17
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
2

Relationship

1
5

Authors

Journals

citations
Cited by 19 publications
(17 citation statements)
references
References 13 publications
0
17
0
Order By: Relevance
“…Although efforts have been made for ensuring fair comparison between systems [ 11 , 35 ], additional publicly available annotated benchmark datasets, used solely for evaluation purposes, could help the field progress and allow for more comparisons across studies, notably on their ability to generalize to new data. In this study, by using the WEB-RADR reference dataset, a publicly available dataset [ 43 ], we identified a number of factors that could explain the poor transferability of the system we developed and of another published system aimed at classifying AE posts. The poor transferability offers a plausible explanation to why, despite almost a decade since the first AE recognition systems in social media have been published, such systems have not been adopted in routine pharmacovigilance practice.…”
Section: Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…Although efforts have been made for ensuring fair comparison between systems [ 11 , 35 ], additional publicly available annotated benchmark datasets, used solely for evaluation purposes, could help the field progress and allow for more comparisons across studies, notably on their ability to generalize to new data. In this study, by using the WEB-RADR reference dataset, a publicly available dataset [ 43 ], we identified a number of factors that could explain the poor transferability of the system we developed and of another published system aimed at classifying AE posts. The poor transferability offers a plausible explanation to why, despite almost a decade since the first AE recognition systems in social media have been published, such systems have not been adopted in routine pharmacovigilance practice.…”
Section: Discussionmentioning
confidence: 99%
“…To our knowledge, this study is the first to present the development of an AE recognition system together with a prospective evaluation of its performance outside of the universe of the data it has been trained on. We perform an external evaluation using a publicly available benchmark dataset manually curated and annotated by members of the WEB-RADR consortium [43]. The dataset is entirely independent from the dataset we used for training our system, which was provided to us by Epidemico, a health informatics company (later acquired by Booz Allen Hamilton) and former WEB-RADR partner.…”
Section: Key Pointsmentioning
confidence: 99%
See 1 more Smart Citation
“…Among tweets that mention medications, which is often a starting point for data collection, tweets that mention ADE are outnumbered 10:1 to 50:1 by tweets that do not contain ADEs. 7 , 8 , 12 , 13 From our preliminary analysis of datasets in shared tasks, the variability in the ratios could be largely attributed to the class of drugs being used for the study. Emerging medications are often promoted by bots as well as mentioned in news articles which overshadow firsthand reports of medication consumption by users.…”
Section: Introductionmentioning
confidence: 99%
“…On the other hand, in the second category are tools used to analyze data that are published exclusively on social networks, and that seek to provide users, expert and inexpert, with sufficient elements to perform an easy and intuitive analysis of the results provided by their methods of identification of issues, polarity, etc. As examples of such tools are Spot, AnaliticPro, and Socialmention, which provide the user with various data visualization schemes that seek to highlight certain indicators allowing the analyst to evaluate and determine the reputation of a particular product or topic within a specific community [12].…”
Section: Related Studiesmentioning
confidence: 99%