Summary: Objectives. Scientific and clinical advances in perinatology and neonatology have enhanced the chances of survival of preterm and very low weight neonates. Infant cry analysis is a suitable noninvasive complementary tool to assess the neurologic state of infants particularly important in the case of preterm neonates. This article aims at exploiting differences between full-term and preterm infant cry with robust automatic acoustical analysis and data mining techniques. Study design. Twenty-two acoustical parameters are estimated in more than 3000 cry units from cry recordings of 28 full-term and 10 preterm newborns. Methods. Feature extraction is performed through the BioVoice dedicated software tool, developed at the Biomedical Engineering Lab, University of Firenze, Italy. Classification and pattern recognition is based on genetic algorithms for the selection of the best attributes. Training is performed comparing four classifiers: Logistic Curve, Multilayer Perceptron, Support Vector Machine, and Random Forest and three different testing options: full training set, 10-fold crossvalidation, and 66% split. Results. Results show that the best feature set is made up by 10 parameters capable to assess differences between preterm and full-term newborns with about 87% of accuracy. Best results are obtained with the Random Forest method (receiver operating characteristic area, 0.94). Conclusions. These 10 cry features might convey important additional information to assist the clinical specialist in the diagnosis and follow-up of possible delays or disorders in the neurologic development due to premature birth in this extremely vulnerable population of patients. The proposed approach is a first step toward an automatic infant cry recognition system for fast and proper identification of risk in preterm babies.
Smartphone technology provides new opportunities for recording standardized voice samples of patients and sending the files by e-mail to the voice laboratory. This drastically improves the collection of baseline data, as used in research on efficiency of voice treatments. However, the basic requirement is the suitability of smartphones for recording and digitizing pathologic voices (mainly characterized by period perturbations and noise) without significant distortion. In this experiment, two smartphones (a very inexpensive one and a high-level one) were tested and compared with direct microphone recordings in a soundproof room. The voice stimuli consisted in synthesized deviant voice samples (median of fundamental frequency: 120 and 200 Hz) with three levels of jitter and three levels of added noise. All voice samples were analyzed using PRAAT software. The results show high correlations between jitter, shimmer, and noise-to-harmonics ratio measured on the recordings via both smartphones, the microphone, and measured directly on the sound files from the synthesizer. Smartphones thus appear adequate for reliable recording and digitizing of pathologic voices.
Egocentric vision (a.k.a. first-person vision -FPV) applications have thrived over the past few years, thanks to the availability of affordable wearable cameras and large annotated datasets. The position of the wearable camera (usually mounted on the head) allows recording exactly what the camera wearers have in front of them, in particular hands and manipulated objects. This intrinsic advantage enables the study of the hands from multiple perspectives: localizing hands and their parts within the images; understanding what actions and activities the hands are involved in; and developing human-computer interfaces that rely on hand gestures. In this survey, we review the literature that focuses on the hands using egocentric vision, categorizing the existing approaches into: localization (where are the hands or part of them?); interpretation (what are the hands doing?); and application (e.g., systems that used egocentric hand cues for solving a specific problem). Moreover, a list of the most prominent datasets with hand-based annotations is provided.
Rapid maxillary expansion causes a slight phonetic change in the acoustical parameters of both consonants and vowels. The two-arm Hyrax caused less speech impairment than the four-arm Hyrax during the treatment.
Summary: Objectives. A large percentage of patients with Parkinson's disease have hypokinetic dysarthria, exhibiting reduced peak velocities of jaw and lips during speech. This limitation implies a reduction of speech intelligibility for such patients. This work aims at testing a cost-effective markerless approach for assessing kinematic parameters of hypokinetic dysarthria. Study Design. Kinematic parameters of the lips are calculated during a syllable repetition task from 14 Parkinsonian patients and 14 age-matched control subjects. Methods. Combining color and depth frames provided by a depth sensor (Microsoft Kinect), we computed the threedimensional coordinates of main facial points. The peak velocities and accelerations of the lower lip during a syllable repetition task are considered to compare the two groups. Results. Results show that Parkinsonian patients exhibit reduced peak velocities of the lower lip, both during the opening and the closing phase of the mouth. In addition, peak values of acceleration are reduced in Parkinsonian patients, although with significant differences only in the opening phase with respect to healthy control subjects. Conclusions. The novel contribution of this work is the implementation of an entirely markerless technique capable to detect signs of hypokinetic dysarthria for the analysis of articulatory movements during speech. Although a large number of Parkinsonian patients have hypokinetic dysarthria, only a small percentage of them undergoes speech therapy to increase their articulatory movements. The system proposed here could be easily implemented in a home environment, thus, increasing the percentage of patients who can perform speech rehabilitation at home.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.