2012
DOI: 10.1007/s10994-012-5296-5
|View full text |Cite
|
Sign up to set email alerts
|

Structured learning with constrained conditional models

Abstract: Making complex decisions in real world problems often involves assigning values to sets of interdependent variables where an expressive dependency structure among these can influence, or even dictate, what assignments are possible. Commonly used models typically ignore expressive dependencies since the traditional way of incorporating non-local dependencies is inefficient and hence leads to expensive training and inference.The contribution of this paper is two-fold. First, this paper presents Constrained Condi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
90
0

Year Published

2013
2013
2020
2020

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 93 publications
(91 citation statements)
references
References 30 publications
1
90
0
Order By: Relevance
“…One particular model, initially proposed by Joachims [12] for binary support vector machines, has given impressive results on problems involving large feature spaces, such as those encountered in text classification and natural language processing. This model has been nicely extended to multi-class classification, ordinal regression and structured output problems [31,5,2]. The key ideas behind these methods are: (a) using the labels of the unlabeled data (Y U ) as extra variables and the associated loss function in training; (b) optimizing the model weight vector (θ) and Y U via alternating optimization steps; (c) using constraints on Y U that come from domain knowledge to effectively guide the training towards good solutions; and (d) employing annealing to avoid getting caught in local minima.…”
Section: Introductionmentioning
confidence: 99%
“…One particular model, initially proposed by Joachims [12] for binary support vector machines, has given impressive results on problems involving large feature spaces, such as those encountered in text classification and natural language processing. This model has been nicely extended to multi-class classification, ordinal regression and structured output problems [31,5,2]. The key ideas behind these methods are: (a) using the labels of the unlabeled data (Y U ) as extra variables and the associated loss function in training; (b) optimizing the model weight vector (θ) and Y U via alternating optimization steps; (c) using constraints on Y U that come from domain knowledge to effectively guide the training towards good solutions; and (d) employing annealing to avoid getting caught in local minima.…”
Section: Introductionmentioning
confidence: 99%
“…We approach this challenge using undirected probabilistic graphical models which integrate coherence constraints over pairs of slots within a schema. Similar techniques have been proposed for the more shallow problems of HMM-based sequence labeling by Chang et al [5] and relation extraction by Lopez de Lacalle & Lapata [12]. In line with the latter approach, we aim at inducing constraint knowledge automatically from training data.…”
Section: Related Workmentioning
confidence: 98%
“…We followed the settings in [3], where the HMM is assumed to be used for identifying the role, such as 'author' and 'title', of each word in a given citation text snippet. The sequential labeling task correspond to finding the Viterbi path of the HMM.…”
Section: Experimental Settingsmentioning
confidence: 99%
“…We trained the HMM with 10 training samples and used 99 samples as test data. Both sets were obtained from [3]. There are 12 types of constraints.…”
Section: Experimental Settingsmentioning
confidence: 99%
See 1 more Smart Citation