2011
DOI: 10.1007/978-3-642-23780-5_20
|View full text |Cite
|
Sign up to set email alerts
|

A Spectral Learning Algorithm for Finite State Transducers

Abstract: Abstract. Finite-State Transducers (FSTs) are a popular tool for modeling paired input-output sequences, and have numerous applications in real-world problems. Most training algorithms for learning FSTs rely on gradient-based or EM optimizations which can be computationally expensive and suffer from local optima issues. Recently, Hsu et al. [13] proposed a spectral method for learning Hidden Markov Models (HMMs) which is based on an Observable Operator Model (OOM) view of HMMs. Following this line of work we p… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
14
0

Year Published

2012
2012
2022
2022

Publication Types

Select...
5
3
1

Relationship

5
4

Authors

Journals

citations
Cited by 14 publications
(14 citation statements)
references
References 22 publications
(43 reference statements)
0
14
0
Order By: Relevance
“…In the last years multiple spectral learning algorithms have been proposed for a wide range of models. Many of these models deal with data whose nature is eminently sequential, like the work of Bailly et al (2009) on WFA, or other works on particular subclasses of WFA like HMM (Hsu et al 2009) and related extensions (Siddiqi et al 2010;Song et al 2010), Predictive State Representations (PSR) (Boots et al 2011), Finite State Transducers (FST) (Balle et al 2011), and Quadratic Weighted Automata (QWA) (Bailly 2011). Besides direct applications of the spectral algorithm to different classes of sequential models, the method has also been combined with convex optimization algorithms in , Balle and Mohri (2012).…”
Section: Related Workmentioning
confidence: 99%
“…In the last years multiple spectral learning algorithms have been proposed for a wide range of models. Many of these models deal with data whose nature is eminently sequential, like the work of Bailly et al (2009) on WFA, or other works on particular subclasses of WFA like HMM (Hsu et al 2009) and related extensions (Siddiqi et al 2010;Song et al 2010), Predictive State Representations (PSR) (Boots et al 2011), Finite State Transducers (FST) (Balle et al 2011), and Quadratic Weighted Automata (QWA) (Bailly 2011). Besides direct applications of the spectral algorithm to different classes of sequential models, the method has also been combined with convex optimization algorithms in , Balle and Mohri (2012).…”
Section: Related Workmentioning
confidence: 99%
“…Recently a number of researchers have developed provably correct algorithms for parameter estimation in latent variable models such as hidden Markov models, topic models, directed graphical models with latent variables, and so on (Hsu et al, 2009;Bailly et al, 2010;Siddiqi et al, 2010;Parikh et al, 2011;Balle et al, 2011;Arora et al, 2013;Dhillon et al, 2012;Anandkumar et al, 2012;Arora et al, 2012;Arora et al, 2013). Many of these algorithms have their roots in spectral methods such as canonical correlation analysis (CCA) (Hotelling, 1936), or higher-order tensor decompositions.…”
Section: Related Workmentioning
confidence: 99%
“…It is also important to note that MLE is not the only option for estimating finite state probabilistic grammars. There has been some recent advances in learning finite state models (HMMs and finite state transducers) by using spectral analysis of matrices which consist of quantities estimated from observations only (Hsu, Kakade, and Zhang 2009;Balle, Quattoni, and Carreras 2011), based on the observable operator models of Jaeger (1999). These algorithms are not prone to local minima, and converge to the correct model as the number of samples increases, but require some assumptions about the underlying model that generates the data.…”
Section: Discussionmentioning
confidence: 99%