“…Fairness Techniques. Prior work in language [4,20,31], computer vision [17,51], and graphs [1,27,40,54] has primarily focused on debiasing models trained on unimodal data and are limited in scope as they only investigate gender bias, racial bias, or their intersections. In particular, their bias mitigation techniques can be broadly categorized into i) pre-processing, which modifies individual input features and labels [6], modifies the weights of the training samples [24], or obfuscates protected attribute information during the training process [59]; ii) in-processing, which uses adversarial techniques to maximize accuracy and reduce bias for a given protected attribute [61], data aug-mentation [1] or adding a bias-aware regularization term to the training objectives [26], and iii) post-processing, which changes the output predictions from predictive models to make them fairer [21,25,44].…”