The Carnegie Mellon University's (CMU) Sphinx framework is increasingly used for the Arabic speech recognition in general and applied to the Holy Quran in particular. Generating the language model includes a tedious task of preparing the transcriptions for all the data. In this paper, we investigate the faulttolerance of the automatically generated language model as compared to a corrected and uncorrected transcription with and without silence tagging. This editing addresses the different repetitions and pauses encountered during recitations. Experiments show that the average difference between the lowest and highest Word Error Rate (WER) for each configuration of the number of Senones is 0.6% when using all files for the training and 1.6% when using 80% of the files for training the language model of 17 chapters of the Holy Quran. Results show that the performance of trained models without any correction can be close to when all required rectifications of transcriptions are performed.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.