Emotion recognition from facial expressions is generally performed in three steps: face detection, features extraction and classification of expressions. The present work focuses on two points: Firstly, a new extraction method is presented based on the geometric approach. This method consists of calculating six distances in order to measure parts of the face that better describe a facial expression. Secondly, an automatic supervised learning method called decision tree is applied on two databases (JAFEE and COHEN), in order to have a facial expressions classifying system with seven possible classes (six basic emotions plus neutral); this system uses as input the six distances previously calculated (using Euclidian ,Manhattan or Minkowski distance) for each face. Our results achieved a recognition rate of 89.20% and 90.61% respectively in JAFFE and COHEN database.
The importance of emotion recognition lies in the role that emotions play in our everyday lives. Emotions have a strong relationship with our behavior. Thence, automatic emotion recognition, is to equip the machine of this human ability to analyze, and to understand the human emotional state, in order to anticipate his intentions from facial expression. In this paper, a new approach is proposed to enhance accuracy of emotion recognition from facial expression, which is based on input features deducted only from fiducial points. The proposed approach consists firstly on extracting 1176 dynamic features from image sequences that represent the proportions of euclidean distances between facial fiducial points in the first frame, and faicial fiducial points in the last frame. Secondly, a feature selection method is used to select only the most relevant features from them. Finally, the selected features are presented to a Neural Network (NN) classifier to classify facial expression input into emotion. The proposed approach has achieved an emotion recognition accuracy of 99% on the CK+ database, 84.7% on the Oulu-CASIA VIS database, and 93.8% on the JAFFE database.
The deployment of 4G/LTE (Long Term Evolution) mobile network has solved the major challenge of high capacities, to build real broadband mobile Internet. This was possible mainly through very strong physical layer and flexible network architecture. However, the bandwidth hungry services have been developed in unprecedented way, such as virtual reality (VR), augmented reality (AR), etc. Furthermore, mobile networks are facing other new services with extremely demand of higher reliability and almost zero-latency performance, like vehicle communications or Internet-of-Vehicles (IoV). Using new radio interface based on massive MIMO, 5G has overcame some of these challenges. In addition, the adoption of software defend networks (SDN) and network function virtualization (NFV) has added a higher degree of flexibility allowing the operators to support very demanding services from different vertical markets. However, network operators are forced to consider a higher level of intelligence in their networks, in order to deeply and accurately learn the operating environment and users behaviors and needs. It is also important to forecast their evolution to build a pro-actively and efficiently (self-) updatable network. In this chapter, we describe the role of artificial intelligence and machine learning in 5G and beyond, to build cost-effective and adaptable performing next generation mobile network. Some practical use cases of AI/ML in network life cycle are discussed.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.