Abstract:Apparent emotional facial expression recognition has attracted a lot of research attention recently. However, the majority of approaches ignore age differences and train a generic model for all ages. In this work, we study the effect of using different age-groups for training apparent emotional facial expression recognition models. To this end, we study Domain Generalisation in the context of apparent emotional facial expression recognition from facial imagery across different age groups. We first compare seve… Show more
“…Demographically speaking, some datasets openly focus on facial expressions of specific demographic groups, such as JAFFE [46], [47] (Japanese women) and iSAFE [48] (Indian people), but for the most part the datasets have not been collected taking diversity into account. Some recent works have already found specific biases around gender, race, and age in both commercial FER systems [49], [50] and research models [51], [52], [53], [54], [55], [56]. From these works, Kim et al [49] focus on the age bias of commercial models in an age-labeled dataset.…”
Section: Facial Expression Recognitionmentioning
confidence: 99%
“…Deuschel et al [55] analyzes biases in the prediction of action units with respect to gender and skin color in two popular datasets. Poyiadzy et al [56] work on age bias in a dataset collected from Internet searches, performing additional work to generate apparent age labels for it. Additionally, some works [52], [54], [56] have explored mitigation strategies applicable to this problem.…”
Demographic biases in source datasets have been shown as one of the causes of unfairness and discrimination in the predictions of Machine Learning models. One of the most prominent types of demographic bias are statistical imbalances in the representation of demographic groups in the datasets. In this paper, we study the measurement of these biases by reviewing the existing metrics, including those that can be borrowed from other disciplines. We develop a taxonomy for the classification of these metrics, providing a practical guide for the selection of appropriate metrics. To illustrate the utility of our framework, and to further understand the practical characteristics of the metrics, we conduct a case study of 20 datasets used in Facial Emotion Recognition (FER), analyzing the biases present in them. Our experimental results show that many metrics are redundant and that a reduced subset of metrics may be sufficient to measure the amount of demographic bias. The paper provides valuable insights for researchers in AI and related fields to mitigate dataset bias and improve the fairness and accuracy of AI models. The code is available at https://github.com/irisdominguez/dataset bias metrics.
“…Demographically speaking, some datasets openly focus on facial expressions of specific demographic groups, such as JAFFE [46], [47] (Japanese women) and iSAFE [48] (Indian people), but for the most part the datasets have not been collected taking diversity into account. Some recent works have already found specific biases around gender, race, and age in both commercial FER systems [49], [50] and research models [51], [52], [53], [54], [55], [56]. From these works, Kim et al [49] focus on the age bias of commercial models in an age-labeled dataset.…”
Section: Facial Expression Recognitionmentioning
confidence: 99%
“…Deuschel et al [55] analyzes biases in the prediction of action units with respect to gender and skin color in two popular datasets. Poyiadzy et al [56] work on age bias in a dataset collected from Internet searches, performing additional work to generate apparent age labels for it. Additionally, some works [52], [54], [56] have explored mitigation strategies applicable to this problem.…”
Demographic biases in source datasets have been shown as one of the causes of unfairness and discrimination in the predictions of Machine Learning models. One of the most prominent types of demographic bias are statistical imbalances in the representation of demographic groups in the datasets. In this paper, we study the measurement of these biases by reviewing the existing metrics, including those that can be borrowed from other disciplines. We develop a taxonomy for the classification of these metrics, providing a practical guide for the selection of appropriate metrics. To illustrate the utility of our framework, and to further understand the practical characteristics of the metrics, we conduct a case study of 20 datasets used in Facial Emotion Recognition (FER), analyzing the biases present in them. Our experimental results show that many metrics are redundant and that a reduced subset of metrics may be sufficient to measure the amount of demographic bias. The paper provides valuable insights for researchers in AI and related fields to mitigate dataset bias and improve the fairness and accuracy of AI models. The code is available at https://github.com/irisdominguez/dataset bias metrics.
“…Some recent works have already found specific biases around gender, race, and age in both commercial FER systems [41], [42] and research models [43], [44], [45], [46], [47], [48]. From these works, Kim et al [41] focus on the age bias of commercial models in an age-labeled dataset.…”
Section: B Facial Expression Recognitionmentioning
confidence: 99%
“…Deuschel et al [47] analyzes biases in the prediction of action units with respect to gender and skin color in two popular datasets. Poyiadzy et al [48] work on age bias in a dataset collected from Internet searches, performing additional work to generate apparent age labels for it. Additionally, some works [44], [46], [48] have started exploring mitigation strategies applicable to this problem.…”
Section: B Facial Expression Recognitionmentioning
confidence: 99%
“…Poyiadzy et al [48] work on age bias in a dataset collected from Internet searches, performing additional work to generate apparent age labels for it. Additionally, some works [44], [46], [48] have started exploring mitigation strategies applicable to this problem. According to the flaws and biases found in previous studies, Hernandez et al [49] propose a set of guidelines to assess and minimize potential risks in the application of FER-based technologies.…”
Section: B Facial Expression Recognitionmentioning
Demographic biases in source datasets have been shown as one of the causes of unfairness and discrimination in the predictions of Machine Learning models. One of the most prominent types of demographic bias are statistical imbalances in the representation of demographic groups in the datasets. In this paper, we study the measurement of these biases by reviewing the existing metrics, including those that can be borrowed from other disciplines. We develop a taxonomy for the classification of these metrics, providing a practical guide for the selection of appropriate metrics. To illustrate the utility of our framework, and to further understand the practical characteristics of the metrics, we conduct a case study of 20 datasets used in Facial Emotion Recognition (FER), analyzing the biases present in them. Our experimental results show that many metrics are redundant and that a reduced subset of metrics may be sufficient to measure the amount of demographic bias. The paper provides valuable insights for researchers in AI and related fields to mitigate dataset bias and improve the fairness and accuracy of AI models. The code is available at https://github.com/irisdominguez/dataset bias metrics.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.