Abstract:We propose a lightly-supervised approach for information extraction, in particular named entity classification, which combines the benefits of traditional bootstrapping, i.e., use of limited annotations and interpretability of extraction patterns, with the robust learning approaches proposed in representation learning. Our algorithm iteratively learns custom embeddings for both the multi-word entities to be extracted and the patterns that match them from a few example entities per category. We demonstrate that… Show more
“…Because the seed entities are sparse comparing to the number of unexpanded entities, the learned Boot-strapTeacher tends to be underfitting. Inspired by Zupon et al (2019), and based on the intuition that the pattern and entity embeddings are similar to their neighbors but dissimilar to their unrelated patterns or entities, we leverage the graph structure as a regularizer to the learning procedure, and let the BootstrapTeacher maximize the following unsupervised learning objective:…”
Section: Model Learningmentioning
confidence: 99%
“…Datasets: We use two datasets, CoNLL and OntoNotes, constructed by Zupon et al (2019). CoNLL is constructed from the CoNLL 2003 shared task dataset (Tjong Kim Sang and De Meulder 2003), which contains 4 entity types.…”
Section: Experiments Experimental Setupmentioning
confidence: 99%
“…OntoNotes is constructed from the OntoNotes datasets (Pradhan et al 2013) without numerical categories, which finally contains 11 entity types. Zupon et al (2019) use the n-grams of the size up to 4 tokens on either side of an entity as the patterns and filter out some patterns. Baselines.…”
Section: Experiments Experimental Setupmentioning
confidence: 99%
“…• Gupta (Gupta and Manning 2014): this method is a classical bootstrapping system that iteratively evaluates and selects patterns, and scores new entities by a learned entity classifier 4 . • Emboot (Zupon et al 2019): this method follows Gupta and Manning (2014), but learns custom embeddings for entities and patterns at each iteration, which are used to guide the entity classifier. • LTB (Yan et al 2019): this method uses the MCTS algorithm to perform lookahead search for estimating delayed feedback and jointly learns an entity similarity function.…”
Section: Experiments Experimental Setupmentioning
confidence: 99%
“…Second, it is hard to exploit the high-order entity-pattern relations for entity set expansion. Due to the lack of extra supervision, previous studies mainly used the entity-pattern matching statistics for entity/pattern evaluation (Riloff and Jones 1999;Curran, Murphy, and Scholz 2007;Shi et al 2014;Zupon et al 2019). However, these methods only exploit the first-order entity-pattern relations, while ignoring the useful information from the high-order relations, which have been proven useful in many information extraction tasks (Riedel et al 2013;Chen et al 2006).…”
Bootstrapping for entity set expansion (ESE) has long been modeled as a multi-step pipelined process. Such a paradigm, unfortunately, often suffers from two main challenges: 1) the entities are expanded in multiple separate steps, which tends to introduce noisy entities and results in the semantic drift problem; 2) it is hard to exploit the high-order entity-pattern relations for entity set expansion. In this paper, we propose an end-to-end bootstrapping neural network for entity set expansion, named BootstrapNet, which models the bootstrapping in an encoder-decoder architecture. In the encoding stage, a graph attention network is used to capture both the first- and the high-order relations between entities and patterns, and encode useful information into their representations. In the decoding stage, the entities are sequentially expanded through a recurrent neural network, which outputs entities at each stage, and its hidden state vectors, representing the target category, are updated at each expansion step. Experimental results demonstrate substantial improvement of our model over previous ESE approaches.
“…Because the seed entities are sparse comparing to the number of unexpanded entities, the learned Boot-strapTeacher tends to be underfitting. Inspired by Zupon et al (2019), and based on the intuition that the pattern and entity embeddings are similar to their neighbors but dissimilar to their unrelated patterns or entities, we leverage the graph structure as a regularizer to the learning procedure, and let the BootstrapTeacher maximize the following unsupervised learning objective:…”
Section: Model Learningmentioning
confidence: 99%
“…Datasets: We use two datasets, CoNLL and OntoNotes, constructed by Zupon et al (2019). CoNLL is constructed from the CoNLL 2003 shared task dataset (Tjong Kim Sang and De Meulder 2003), which contains 4 entity types.…”
Section: Experiments Experimental Setupmentioning
confidence: 99%
“…OntoNotes is constructed from the OntoNotes datasets (Pradhan et al 2013) without numerical categories, which finally contains 11 entity types. Zupon et al (2019) use the n-grams of the size up to 4 tokens on either side of an entity as the patterns and filter out some patterns. Baselines.…”
Section: Experiments Experimental Setupmentioning
confidence: 99%
“…• Gupta (Gupta and Manning 2014): this method is a classical bootstrapping system that iteratively evaluates and selects patterns, and scores new entities by a learned entity classifier 4 . • Emboot (Zupon et al 2019): this method follows Gupta and Manning (2014), but learns custom embeddings for entities and patterns at each iteration, which are used to guide the entity classifier. • LTB (Yan et al 2019): this method uses the MCTS algorithm to perform lookahead search for estimating delayed feedback and jointly learns an entity similarity function.…”
Section: Experiments Experimental Setupmentioning
confidence: 99%
“…Second, it is hard to exploit the high-order entity-pattern relations for entity set expansion. Due to the lack of extra supervision, previous studies mainly used the entity-pattern matching statistics for entity/pattern evaluation (Riloff and Jones 1999;Curran, Murphy, and Scholz 2007;Shi et al 2014;Zupon et al 2019). However, these methods only exploit the first-order entity-pattern relations, while ignoring the useful information from the high-order relations, which have been proven useful in many information extraction tasks (Riedel et al 2013;Chen et al 2006).…”
Bootstrapping for entity set expansion (ESE) has long been modeled as a multi-step pipelined process. Such a paradigm, unfortunately, often suffers from two main challenges: 1) the entities are expanded in multiple separate steps, which tends to introduce noisy entities and results in the semantic drift problem; 2) it is hard to exploit the high-order entity-pattern relations for entity set expansion. In this paper, we propose an end-to-end bootstrapping neural network for entity set expansion, named BootstrapNet, which models the bootstrapping in an encoder-decoder architecture. In the encoding stage, a graph attention network is used to capture both the first- and the high-order relations between entities and patterns, and encode useful information into their representations. In the decoding stage, the entities are sequentially expanded through a recurrent neural network, which outputs entities at each stage, and its hidden state vectors, representing the target category, are updated at each expansion step. Experimental results demonstrate substantial improvement of our model over previous ESE approaches.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.