2008
DOI: 10.1007/s11390-008-9157-4
|View full text |Cite
|
Sign up to set email alerts
|

Scaling Conditional Random Fields by One-Against-the-Other Decomposition

Abstract: As a powerful sequence labeling model, conditional random fields (CRFs) have had successful applications in many natural language processing (NLP) tasks. However, the high complexity of CRFs training only allows a very small tag (or label) set, because the training becomes intractable as the tag set enlarges. This paper proposes an improved decomposed training and joint decoding algorithm for CRF learning. Instead of training a single CRF model for all tags, it trains a binary sub-CRF independently for each ta… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2010
2010
2011
2011

Publication Types

Select...
3

Relationship

1
2

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 11 publications
0
3
0
Order By: Relevance
“…CRFs [4,9] are the state-of-the-art approaches in information extraction taking the sequence characteristics to do better labeling, which have been widely used in many fields, such as NLP tasks, IE. To better incorporate the two-dimensional neighborhood dependencies, Zhu et al [1] propose a two-dimensional Conditional Random Fields (2DCRFs) model for semantic annotation of Web objects.…”
Section: Related Workmentioning
confidence: 99%
“…CRFs [4,9] are the state-of-the-art approaches in information extraction taking the sequence characteristics to do better labeling, which have been widely used in many fields, such as NLP tasks, IE. To better incorporate the two-dimensional neighborhood dependencies, Zhu et al [1] propose a two-dimensional Conditional Random Fields (2DCRFs) model for semantic annotation of Web objects.…”
Section: Related Workmentioning
confidence: 99%
“…The other is the conditional random fields (CRFs) model [23] for supervised segmentation via character tagging, conventionally trained only on a pre-segmented corpus. The latter is a state-of-the-art approach that has set new performance records in the field, as illustrated in [52,55], although its efficiency is yet to be further enhanced by various means [58,59]. All scores given by the goodness measures are discretized in the same way for use as feature values in the CRFs model.…”
Section: Introductionmentioning
confidence: 99%
“…They adopted the construction features and time features of new words, but the experimental result was not good. Peng [10] regarded the process of Chinese word segmentation and new word identification as a unified step using the Conditional Random Field (CRF) model [11][12] , the method only detects new words and does not assign POS tags to new words. Peng [10] proved that the character-based model performs better than the wordbased model in new word identification.…”
Section: Introductionmentioning
confidence: 99%