“…Teams Overall A B C s2 end2end (Ammar et al, 2017) 0.43 0.55 0.44 0.28 TIAL UW 0.42 0.56 0.44 TTI COIN (Tsujimura et al, 2017) 0.38 0.5 0.39 0.21 PKU ICL (Wang and Li, 2017) 0.37 0.51 0.38 0.19 NTNU-1 0.33 0.47 0.34 0.2 WING-NUS (Prasad and Kan, 2017) 0.27 0.46 0.33 0.04 Know-Center (Kern et al, 2017) 0.27 0.39 0.28 SZTE-NLP (Berend, 2017) 0.26 0.35 0.28 NTNU (Lee et al, 2017b) 0.23 0.3 0.24 0.08 LABDA (Flores et al, 2017) 0.04 0.08 0.04 upper bound 0.84 0.85 0.85 0.77 random 0.00 0.03 0.01 0.00 former is surprising, as keyphrases are with an overwhelming majority noun phrases, the latter not as much, many keyphrases only appear once in the dataset (see Table 1). GMBUAP further tried using empirical rules obtained by observing the training data for Subtask A, and a Naive Bayes classifier trained on provided training data for Subtask B.…”