2005
DOI: 10.1186/1471-2105-6-s1-s12
|View full text |Cite
|
Sign up to set email alerts
|

Data preparation and interannotator agreement: BioCreAtIvE Task 1B

Abstract: Background: We prepared and evaluated training and test materials for an assessment of text mining methods in molecular biology. The goal of the assessment was to evaluate the ability of automated systems to generate a list of unique gene identifiers from PubMed abstracts for the three model organisms Fly, Mouse, and Yeast. This paper describes the preparation and evaluation of answer keys for training and testing. These consisted of lists of normalized gene names found in the abstracts, generated by adapting … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
29
0

Year Published

2005
2005
2015
2015

Publication Types

Select...
4
2

Relationship

1
5

Authors

Journals

citations
Cited by 27 publications
(29 citation statements)
references
References 7 publications
0
29
0
Order By: Relevance
“…This meant that we had to edit the gene lists to make them correspond to genes mentioned in the abstract, rather than all the genes curated in the full text article. We developed a procedure to automatically remove genes not found in the abstract and were able to provide a large quantity of "noisy" training data for the three organisms, together with small collections of carefully corrected development and test data [11]. We estimated the quality of the noisy training data for the three organisms.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…This meant that we had to edit the gene lists to make them correspond to genes mentioned in the abstract, rather than all the genes curated in the full text article. We developed a procedure to automatically remove genes not found in the abstract and were able to provide a large quantity of "noisy" training data for the three organisms, together with small collections of carefully corrected development and test data [11]. We estimated the quality of the noisy training data for the three organisms.…”
Section: Resultsmentioning
confidence: 99%
“…There are 6 papers for task 1b, including an overview [10], an article describing preparation of the test sets and inter-annotator agreement experiments [11], and four articles describing systems and results for task 1b [9,12-14]. …”
Section: Introductionmentioning
confidence: 99%
“…Whenever new annotators joined the project, they had to be trained using previously annotated examples and follow the guideline. Colosimo et al [5] and Tanabe et al [28] also conduct corpus annotation in the biology domain and conclude that clear annotation guidelines are important, and the annotations should be validated by proper interannotator-agreement experiments.…”
Section: How To Annotate Properly: What Have We Learnt?mentioning
confidence: 99%
“…3.6. Ideally, each topic would be judged by at least three judges and use a consensus-driven process which would provide maximum consistency throughout the entire process of gold standard creation (Hripcsak and Wilcox 2002;Colosimo et al 2005). The time scale of TREC and the resources required for a consensus approach made this impractical.…”
Section: Factors Influencing Inter-annotator Agreementmentioning
confidence: 99%
“…For the 2002 KDD cup and year 1 of the Genomics track, assessment was performed by using existing curation in FlyBase and Entrez Gene, respectively (Hersh and Bhupatiraju 2003;Yeh et al 2003). Task 1a of the first BioCreAtIvE challenge took a multipronged approach by developing guidelines for their three assessors, performing three-way comparisons of replicated judgments, and using pooled results from participants to uncover potential false negatives and false positives (Colosimo et al 2005). For judging duties that went beyond relevance assessment, the TREC Genomics track employed guidelines, training sessions, scaled-down ''mini-topics'', and moderated assessment by an experienced former judge.…”
Section: Factors Influencing Inter-annotator Agreementmentioning
confidence: 99%