Human age information is present in the face image and speech, but most age estimation methods focus on single-modal data only. Although multi-modal approaches have achieved promising performance in other fields by leveraging complementarity between modalities, the development of age estimation has hardly been more inspired by them because of the problem of modality missing in age data. In this paper, we propose an Embedding-Regularized Double Branches Fusion (ERDBF) framework that can simultaneously handle single-modal and multi-modal age estimation. To address the modality missing, we design a double branches fusion network which consists of an embedding regularization module and an information interaction module. The former aims to enhance the representational capacity of single-modal features. The latter learns inter-modal complementary information. Through their collaboration in the fusion process, the network can extract discriminative age representations, and even if a certain modality is missing, robust age estimation can be achieved using enhanced single-modal information. To our best knowledge, the proposed framework is the first deep learning-based multi-modal age estimation framework that can easily utilize the achievements made in the single-modal and have broader applications. Experimental results show that our best method achieves state-of-the-art results on AgeVoxCeleb. The code is available at https://github.com/Daretowin/ERDBF.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.