2015
DOI: 10.17743/jaes.2015.0068
|View full text |Cite
|
Sign up to set email alerts
|

Perceived Audio Quality of Sounds Degraded by Nonlinear Distortions and Single-Ended Assessment Using HASQI

Abstract: For field recordings and user generated content recorded on phones, tablets, and other mobile devices nonlinear distortions caused by clipping and limiting at pre-amplification stages, and dynamic range control (DRC) are common causes of poor audio quality. A single-ended method to detect these distortions and predict perceived degradation in speech, music, and soundscapes has been developed. This was done by training an ensemble of decision trees. During training, both clean and distorted audio was available … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 11 publications
(5 citation statements)
references
References 18 publications
(19 reference statements)
0
5
0
Order By: Relevance
“…Both metrics incorporate a model of the auditory periphery that includes auditory frequency analysis that depends on signal level and audiogram, auditory dynamic-range compression, the shift in auditory threshold that corresponds to the hearing loss, and neural firing-rate adaptation. The metrics have been validated for a wide range of hearing-aid processing algorithms and signal degradation conditions, including frequency-response shaping, wide dynamic-range compression (WDRC), noise suppression, feedback cancellation, frequency lowering, additive noise and babble, and nonlinear distortion such as amplitude quantization and peak clipping (Kates and Arehart, 2010; Houben et al ., 2011; Kressner et al ., 2013; Suelzle et al ., 2013; Kates and Arehart, 2014a; Kates and Arehart, 2014b; Huber et al ., 2014; Falk et al ., 2015; Kendrick et al ., 2015, Van Kuyk et al ., 2017).…”
Section: Introductionmentioning
confidence: 99%
“…Both metrics incorporate a model of the auditory periphery that includes auditory frequency analysis that depends on signal level and audiogram, auditory dynamic-range compression, the shift in auditory threshold that corresponds to the hearing loss, and neural firing-rate adaptation. The metrics have been validated for a wide range of hearing-aid processing algorithms and signal degradation conditions, including frequency-response shaping, wide dynamic-range compression (WDRC), noise suppression, feedback cancellation, frequency lowering, additive noise and babble, and nonlinear distortion such as amplitude quantization and peak clipping (Kates and Arehart, 2010; Houben et al ., 2011; Kressner et al ., 2013; Suelzle et al ., 2013; Kates and Arehart, 2014a; Kates and Arehart, 2014b; Huber et al ., 2014; Falk et al ., 2015; Kendrick et al ., 2015, Van Kuyk et al ., 2017).…”
Section: Introductionmentioning
confidence: 99%
“…where x + (n) and x À (n) are mixed signals, s(n) is the speech signal and w(n) is the noise signal, all in the discrete time domain, i.e., as function of the sampled time n. We used Gaussian noise with a lower band limit of 20 Hz and an upper band limit of 8 kHz to mask speech sampled at 16 kHz. With equation (1) we calculated one estimated speech signal for each mixed signal of equation (2). By summing the two estimated speech signals, the residual noises cancel out and we obtain the clean speech signal with musical noise distortions.…”
Section: Adding Musical Noise To a Signalmentioning
confidence: 99%
“…Nonspeech audio and speech signal distortions appear in many communication systems, e.g., in telephones, headphones, and hearing aids. The causes of these distortions are manifold, e.g., bandwidth limitations, audio codecs, internal noise, dynamic range compression, or noise reduction artifacts [1][2][3][4]. Many psychoacoustical experiments rely on a controllable simulation of these distortions to examine human perception of distorted audio signals.…”
Section: Introductionmentioning
confidence: 99%
“…In addition, HASQI has been validated by several perceptual quality experiments. 3,[13][14][15] HASQI is therefore sensitive to changes in the speech spectrum introduced by acoustic feedback, whistling or ringing in the HA, and any nonlinear distortion introduced by the feedback-cancellation processing.…”
Section: Hasqimentioning
confidence: 99%