The expression of "naming," "commanding," "angry," "frightened," "pleading," "astonished," "satisfied," "admiring," "scornful," and "sad" was with the word [saara] spoken by 12 subjects. Using the same connotations, the 120 utterances were categorized by 73 listeners. Most samples were agreed on by 50%-99% of the judges. Most samples of "astonished," "angry," "frightened," and "commanding" were judged as intended, while "pleading" was often confused with "sad," and "content" with "admiring." Acoustic differences between the categories were examined for F0, duration, and sound pressure; spectral features of [aa] were visualized with the self-organizing map of Kohonen. Most intraspeaker variation of mean F0, FO range, sound pressure, and duration took place during the [aa] segment. Peak sound pressure, mean FO, and spectral energy distribution of [aa] differentiated among "commanding," "angry," "frightened," "naming," and "sad.". Specific intonations of the [aa] segment were encountered for "astonished," "scornful," and "pleading.". The best-conveyed "admiring" samples were distinguished from "content" by spectral cues for a breathy voice quality.
To obtain a perceptual reference for acoustic feature selection, 94 male and 124 female voices were categorized using the ratings of 6 clinicians on visual analog scales for pathology, roughness, breathiness, strain, asthenia, and pitch. Partial correlations showed that breathiness and roughness were the main determinants of pathology. The six-dimensional ratings (the six median scores for each voice) were categorized with the aid of the Sammon map and the self-organizing map. The five categories created differed with respect to the breathiness/roughness ratio and the degree of pathology.
The self-organizing map, a neural network algorithm of Kohonen, was used for the detection of coarticulatory variation of fricative [s] preceding vowels [a:], [i:], and [u:]. The results were compared with the psychoacoustic classification of the same samples to find out whether the map had extracted perceptually meaningful features of [s]. The map distinguished samples of [s] in front of [u:] from those in front of [a:] or [i:] throughout the fricative duration. Samples of [s] preceding [a:] and [i:] were distinguished from each other only just before (about 40 ms) the vowel onset. The results agreed with the perceptual classifications. Most judgments (82%) of [s] in front of [u:] were correct, and this variant of [s] was recognized from the first and second halves of segmented fricatives equally well. Samples of [s] in front of [a:] and [i:] were distinguished from each other less accurately. When halves of segmented [s] were perceptually judged, the differentiation between the following [a] and [i] was possible only on the basis of the second half of the fricative. The results demonstrate that the self-organizing map is a useful tool for the extraction of intersubject regularities in speech spectra. The map also provides an easily understandable, on-line, visualization of speech that can be used as feedback in therapy and education.
Acoustic differences between samples of [i], [u], and [a] uttered in nose-open and nose-obstructed condition were studied in 6 women with isolated cleft palate and pathological nasalance scores and 9 healthy women with normal nasalance scores. The speech samples were depicted by 14-component vocal tract area feature vectors obtained by linear prediction and the differences between the samples were studied with a self organized feature map. Each location on the map corresponds to a certain signal pattern, neighboring locations to similar patterns. The group of healthy subjects differed from the patients for the vowels [i] and [u] but not for [a]. In the patients the nose obstruction induced a significant change in the location of these vowel samples on the map. In healthy subjects no such changes were detected. The result agreed with perceived differences between the subjects.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.