Speech Emotion Classification (SEC) relies heavily on the quality of feature extraction and selection from the speech signal. Improvement on this to enhance the classification of emotion had attracted significant attention from researchers. Many primitives and algorithmic solutions for efficient SEC with minimum cost have been proposed; however, the accuracy and performance of these methods have not yet attained a satisfactory point. In this work, we proposed a novel deep transfer learning approach with distinctive emotional rich feature selection techniques for speech emotion classification. We adopt mel-spectrogram extracted from speech signal as the input to our deep convolutional neural network for efficient feature extraction. We froze 19 layers of our pretrained convolutional neural network from re-training to increase efficiency and minimize computational cost. One flattened layer and two dense layers were used. A ReLu activation function was used at the last layer of our feature extraction segment. To prevent misclassification and reduce feature dimensionality, we employed the Neighborhood Component Analysis (NCA) feature selection algorithm for picking out the most relevant features before the actual classification of emotion. Support Vector Machine (SVM) and Multi-Layer Perceptron (MLP) classifiers were utilized at the topmost layer of our model. Two popular datasets for speech emotion classification tasks were used, which are: Berling Emotional Speech Database (EMO-DB), and Toronto English Speech Set (TESS), and a combination of EMO-DB with TESS was used in our experiment. We obtained a state-of-the-art result with an accuracy rate of 94.3%, 100% specificity on EMO-DB, and 97.2%, 99.80% on TESS datasets, respectively. The performance of our proposed method outperformed some recent work in SEC after assessment on the three datasets.
Speech emotion classification (SEC) has gained the utmost height and occupied a conspicuous position within the research community in recent times. Its vital role in Human–Computer Interaction (HCI) and affective computing cannot be overemphasized. Many primitive algorithmic solutions and deep neural network (DNN) models have been proposed for efficient recognition of emotion from speech however, the suitability of these methods to accurately classify emotion from speech with multi-lingual background and other factors that impede efficient classification of emotion is still demanding critical consideration. This study proposed an attention-based network with a pre-trained convolutional neural network and regularized neighbourhood component analysis (RNCA) feature selection techniques for improved classification of speech emotion. The attention model has proven to be successful in many sequence-based and time-series tasks. An extensive experiment was carried out using three major classifiers (SVM, MLP and Random Forest) on a publicly available TESS (Toronto English Speech Sentence) dataset. The result of our proposed model (Attention-based DCNN+RNCA+RF) achieved 97.8% classification accuracy and yielded a 3.27% improved performance, which outperforms state-of-the-art SEC approaches. Our model evaluation revealed the consistency of attention mechanism and feature selection with human behavioural patterns in classifying emotion from auditory speech.
Accurate interpretation of chest radiographs outcome in epidemiological studies facilitates the process of correctly identifying chest-related or respiratory diseases. Despite the fact that radiological results have been used in the past and is being continuously used for diagnosis of pneumonia and other respiratory diseases, there abounds much variability in the interpretation of chest radiographs. This variability often leads to wrong diagnosis due to the fact that chest diseases often have common symptoms. Moreover, there is no single reliable test that can identify the symptoms of pneumonia. Therefore, this paper presents a standardized approach using convolutional neural network (CNN) and transfer learning technique for identifying pneumonia from chest radiographs that ensure accurate diagnosis and assist physicians in making precise prescriptions for the treatment of pneumonia. A training set consisting of 5,232 optical coherence tomography and chest X-ray images dataset from Mendelev public database was used for this research and the performance evaluation of the model developed on the test set yielded 88.14% accuracy, 90% precision, 85% recall and F1 score of 0.87.
Learning is any technique that in living creatures prompts never-ending limit change and which isn't solely a direct result of characteristic advancement of developing. The complexity associated with learning and the fact that it start from birth till death makes it a cumbersome procedure. It incorporates certainly more than reasoning: the whole character - resources, feelings, impulse, values and will. Many conventional approaches fail to inculcate the above parameters which increase the cumbersomeness of learning coupled with problems of assimilation. If we don't have the will to learn, we won't learn and if we have learned, we are truly changed by one way or another. The focus of this paper is to propose an architecture that was designed with special emphasis on enhancing adaptive elearning. This architecture uses the learning style of learner to produce learning contents peculiar to such learner and as such difficulties associated with comprehension is totally aborted and thereby making learning easier.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.