2016
DOI: 10.1016/j.neucom.2015.07.135
|View full text |Cite
|
Sign up to set email alerts
|

Convolutional feature learning and Hybrid CNN-HMM for scene number recognition

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
18
0

Year Published

2017
2017
2022
2022

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 55 publications
(20 citation statements)
references
References 41 publications
1
18
0
Order By: Relevance
“…The output layer of the CNN is using the softmax function to compute the posterior probability P ( x i | o i ) of class x i given the input observations. It has been shown that by using Baye's rule, the emission probability can be computed by the scaled likelihood: Pfalse(oifalse|xifalse)=Pfalse(xifalse|oifalse)Pfalse(xifalse), where P ( x i ) is the prior probability of class x at state i , and it is computed by counting the number of state class in the training examples. Pfalse(xifalse|oifalse)=yx=expfalse(zxfalse)xexpfalse(zxfalse), where z x is the output neurons from the previous layer multiplied by the weights.…”
Section: Proposed Methodsmentioning
confidence: 99%
“…The output layer of the CNN is using the softmax function to compute the posterior probability P ( x i | o i ) of class x i given the input observations. It has been shown that by using Baye's rule, the emission probability can be computed by the scaled likelihood: Pfalse(oifalse|xifalse)=Pfalse(xifalse|oifalse)Pfalse(xifalse), where P ( x i ) is the prior probability of class x at state i , and it is computed by counting the number of state class in the training examples. Pfalse(xifalse|oifalse)=yx=expfalse(zxfalse)xexpfalse(zxfalse), where z x is the output neurons from the previous layer multiplied by the weights.…”
Section: Proposed Methodsmentioning
confidence: 99%
“…The output holds scores of the classes [8]- [10]. For creating CNN classifiers, a few distinct types of layers are commonly used [2], [3], [8], [9]. They are convolutional layers (ConvL), rectified linear unit layer (ReLU), average pooling layer (AvPL), max pooling layer (MaxPL), fully connected layer (FCL), softmax layer (SML) and dropout layer (DOL) [1]- [3], [8], [10], [11].…”
Section: An Open Problem Of Setting Hyperparametersmentioning
confidence: 99%
“…Segmenting and recognizing the characters go hand in hand. With a goal to amalgamate segmentation and recognition, we will follow the proposed theory of hybrid CNN-HMM [18] where the model undergoes sliding window mechanism to extract a series of frames. Feature descriptors like SIFT, HOG and LBP are widely used in vision problems.…”
Section: International Journal Of Computer Applications (0975 -8887) mentioning
confidence: 99%