ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2020
DOI: 10.1109/icassp40776.2020.9054569
|View full text |Cite
|
Sign up to set email alerts
|

Structural Sparsification for Far-Field Speaker Recognition with Intel® Gna

Abstract: Recently, deep neural networks (DNN) have been widely used in speaker recognition area. In order to achieve fast response time and high accuracy, the requirements for hardware resources increase rapidly. However, as the speaker recognition application is often implemented on mobile devices, it is necessary to maintain a low computational cost while keeping high accuracy in far-field condition. In this paper, we apply structural sparsification on time-delay neural networks (TDNN) to remove redundant structures … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2020
2020
2020
2020

Publication Types

Select...
1

Relationship

1
0

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 24 publications
(26 reference statements)
0
1
0
Order By: Relevance
“…Sharan et al [21] explore random projection for low-rank tensor factorization and describe the use on gene expression and EEG time series data. Zhang et al [22] apply structural sparsification on Time-Delay Neural Networks (TDNN) to remove redundant structures. Alternative approaches are subject to our further research, e.g., binary neural networks as successfully applied to natural language understanding [23].…”
Section: Introductionmentioning
confidence: 99%
“…Sharan et al [21] explore random projection for low-rank tensor factorization and describe the use on gene expression and EEG time series data. Zhang et al [22] apply structural sparsification on Time-Delay Neural Networks (TDNN) to remove redundant structures. Alternative approaches are subject to our further research, e.g., binary neural networks as successfully applied to natural language understanding [23].…”
Section: Introductionmentioning
confidence: 99%