Fair Representation Learning with Unreliable Labels

Publication Type:
Conference Proceeding
Citation:
Proceedings of Machine Learning Research, 2023, 206, pp. 4655-4667
Issue Date:
2023-01-01
Full metadata record
In learning with fairness, for every instance, its label can be systematically flipped to another class due to the practitioner's prejudice, namely, label bias. The existing well-studied fair representation learning methods focus on removing the dependency between the sensitive factors and the input data, but do not address how the representations retain useful information when the labels are unreliable. In fact, we find that the learned representations become random or degenerated when the instance is contaminated by label bias. To alleviate this issue, we investigate the problem of learning fair representations that are independent of the sensitive factors while retaining the task-relevant information given only access to unreliable labels. Our model disentangles the dependency between fair representations and sensitive factors in the latent space. To remove the reliance between the labels and sensitive factors, we incorporate an additional penalty based on mutual information. The learned purged fair representations can then be used in any downstream processing. We demonstrate the superiority of our method over previous works through multiple experiments on both synthetic and real-world datasets.
Please use this identifier to cite or link to this item: