2010 IEEE International Conference on Acoustics, Speech and Signal Processing 2010
DOI: 10.1109/icassp.2010.5495735
|View full text |Cite
|
Sign up to set email alerts
|

An extension of Separable Lattice 2-D HMMS for rotational data variations

Abstract: This paper proposes a new generative model which can deal with rotational data variations by extending Separable Lattice 2-D HMMs (SL2D-HMMs). In image recognition, geometrical variations such as size, location and rotation degrade the performance, therefore normalization is required. SL2D-HMMs can perform an elastic matching in both horizontal and vertical directions; this makes it possible to model invariances to size and location. To deal with rotational variations, we introduce additional HMM states which … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2012
2012
2018
2018

Publication Types

Select...
3
2

Relationship

3
2

Authors

Journals

citations
Cited by 5 publications
(3 citation statements)
references
References 4 publications
0
3
0
Order By: Relevance
“…One of the advantages of SL-HMMs over CNNs is explicit modeling of the generative process, which can represent geometric variations over an entire image. Furthermore, some extensions to structures representing typical geometric variations that are seen in many image recognition tasks have already been proposed, e.g., a structure for rotational variations [18], a structure with multiple horizontal and vertical Markov chains [19], and explicit state duration modeling [20]. By selecting an appropriate model structure reflecting the data generation process for a target task, human knowledge can effectively be utilized as prior information, and this makes it possible to construct models with a small amount of training data.…”
Section: Introductionmentioning
confidence: 99%
“…One of the advantages of SL-HMMs over CNNs is explicit modeling of the generative process, which can represent geometric variations over an entire image. Furthermore, some extensions to structures representing typical geometric variations that are seen in many image recognition tasks have already been proposed, e.g., a structure for rotational variations [18], a structure with multiple horizontal and vertical Markov chains [19], and explicit state duration modeling [20]. By selecting an appropriate model structure reflecting the data generation process for a target task, human knowledge can effectively be utilized as prior information, and this makes it possible to construct models with a small amount of training data.…”
Section: Introductionmentioning
confidence: 99%
“…To cope with this problem, the training algorithm for SL2D-HMMs using the variational EM algorithm were derived in [8], where the loglikelihood can be approximated by the variational lower bound. Although some extensions of SL2D-HMMs have been proposed, e.g., a structure for rotational variations [10], explicit state duration modeling [11], and a structure with multiple horizontal/vertical Markov chains [12], this paper uses an original form of SL2D-HMMs.…”
Section: Separable Lattice 2-d Hmmsmentioning
confidence: 99%
“…One of advantages of SLHMMs against CNNs is explicit modeling of generative process which can represent geometric variations over an entire image. Furthermore, some extensions to structures representing typical geometric variations which are seen in many Copyright c 2016 The Institute of Electronics, Information and Communication Engineers image recognition tasks have already been proposed, e.g., a structure for rotational variations [16], a structure with multiple horizontal and vertical Markov chains [17], explicit state duration modeling [18], trajectory modeling [19], and integration SL-HMMs and factor analyzers [20]. By selecting an appropriate model structures reflecting data generation process for a target task, human knowledge can effectively be utilized as prior information and this makes it possible to construct classifiers with a small amount of training data.…”
Section: Introductionmentioning
confidence: 99%