2019 4th International Conference on Information Technology (InCIT) 2019
DOI: 10.1109/incit.2019.8912049
|View full text |Cite
|
Sign up to set email alerts
|

Audio-Visual Speech Recognition System Using Recurrent Neural Network

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 8 publications
(2 citation statements)
references
References 24 publications
0
2
0
Order By: Relevance
“…In lip-reading studies, it has been noted that some research uses its own datasets, albeit a small number (Lu & Yan, 2020;Goh et al 2019). Generally, studies are conducted on commonly used datasets (Petridis et al 2020;Mesbah et al 2019), known to be OuluVS2 (Anina et al 2015), AvLetters (Matthews et al 2002) and GRID.…”
Section: Datasetsmentioning
confidence: 99%
“…In lip-reading studies, it has been noted that some research uses its own datasets, albeit a small number (Lu & Yan, 2020;Goh et al 2019). Generally, studies are conducted on commonly used datasets (Petridis et al 2020;Mesbah et al 2019), known to be OuluVS2 (Anina et al 2015), AvLetters (Matthews et al 2002) and GRID.…”
Section: Datasetsmentioning
confidence: 99%
“…The audio features are of many kinds. The three of them used in [18] are LPC,PLP, MFCC. The study shows that the MFCC has the highest accuracy of about 94.6% for Hindi Language in noiseless environment.…”
Section: Introductionmentioning
confidence: 99%