2014 43rd International Conference on Parallel Processing Workshops 2014
DOI: 10.1109/icppw.2014.59
|View full text |Cite
|
Sign up to set email alerts
|

GPU-Accelerated HMM for Speech Recognition

Abstract: Speech recognition is used in a wide range of applications and devices such as mobile phones, in-car entertainment systems and web-based services. Hidden Markov Models (HMMs) is one of the most popular algorithmic approaches applied in speech recognition. Training and testing a HMM is computationally intensive and time-consuming. Running multiple applications concurrently with speech recognition could overwhelm the compute resources, and introduce unwanted delays in the speech processing, eventually dropping w… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
7
0

Year Published

2016
2016
2021
2021

Publication Types

Select...
6
2

Relationship

1
7

Authors

Journals

citations
Cited by 12 publications
(7 citation statements)
references
References 14 publications
(10 reference statements)
0
7
0
Order By: Relevance
“…For example, [31], proposes a new distributed multidimensional HMM (DHMM) for multi-object trajectory interaction modeling, the results show superior performance and greater accuracy of the proposed distributed 2D HMM. In [32], the authors present a parallelized HMM to accelerate isolated words speech recognition. Another work of [33] presents a GPU implementation in which they proposed a C and Cuda implementation for the forward, Viterbi and BW algorithms.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…For example, [31], proposes a new distributed multidimensional HMM (DHMM) for multi-object trajectory interaction modeling, the results show superior performance and greater accuracy of the proposed distributed 2D HMM. In [32], the authors present a parallelized HMM to accelerate isolated words speech recognition. Another work of [33] presents a GPU implementation in which they proposed a C and Cuda implementation for the forward, Viterbi and BW algorithms.…”
Section: Related Workmentioning
confidence: 99%
“…Since both the serial forward and the proposed parallel version in each paper were executed using the same dataset with the same parameters, we compute the relative speedup between the two in each case and compare it over the other versions. Table IV shows the result of average relative speedup comparison of ParaDist-Forward algorithm compared to those of [32], [33], [34], [35], [36], [37] and [38]. The results show that the speedup of the proposed model has the best results compare to the benchmark models.…”
Section: B Speedupmentioning
confidence: 99%
“…HMM computation in double precision can be treated as numerically stable. However, in acceleration systems (also GPUs as in [31][32]), where single and half-precision computations are widely used, the length constraints have to be applied to the observation sequence in order to ensure numerical stability.…”
Section: Remarkmentioning
confidence: 99%
“…The efficiency is noticeable for large number of states and iterations. Yu et al [29] achieved 9.2x and 7.9x speedup during the training and testing stages, when used as a Speech Recognition platform for real-time applications. The performance can be limited due to the hardware's memory bandwidth and availability of resources.…”
Section: Related Workmentioning
confidence: 99%
“…The performance can be limited due to the hardware's memory bandwidth and availability of resources. Yu et al [29] states that the GPU version outperforms a single threaded CPU version for internal states greater than 256 for the Forward Algorithm.…”
Section: Related Workmentioning
confidence: 99%