Abstract:
This work studies fair generative models. We reveal and quantify the biases in state-of-the-art (SOTA) GANs w.r.t. different sensitive attributes. To address the biases, ...Show MoreMetadata
Abstract:
This work studies fair generative models. We reveal and quantify the biases in state-of-the-art (SOTA) GANs w.r.t. different sensitive attributes. To address the biases, our main contribution is to propose novel methods to learn fair generative models via transfer learning. Specifically, first, we propose FairTL where we pre-train the generative model with a large biased dataset, then adapt the model using a small fair reference dataset. Second, to further improve sample diversity, we propose FairTL++, containing two additional innovations: 1) aligned feature adaptation, which preserves learned general knowledge while improving fairness by adapting only sensitive attribute-specific parameters, 2) multiple feedback discrimination, which introduces a frozen discriminator for quality feedback and another evolving discriminator for fairness feedback. Taking one step further, we consider an alternative challenging and practical setup. Here, only a pre-trained model is available but the dataset used to pre-train the model is inaccessible. We remark that previous work requires access to large, biased datasets and cannot handle this setup. Extensive experimental results show that FairTL and FairTL++ achieve state-of-the-art performance in quality, diversity and fairness in both setups.
Published in: IEEE Journal of Selected Topics in Signal Processing ( Volume: 18, Issue: 2, March 2024)
Funding Agency:
Keywords assist with retrieval of results and provide a means to discovering other relevant content. Learn more.
- IEEE Keywords
- Index Terms
- Transfer Learning ,
- Transfer Learning Approach ,
- Deep Generative Models ,
- Bias Mitigation ,
- Large Datasets ,
- Small Datasets ,
- Quality Performance ,
- Generative Adversarial Networks ,
- Quality Of Feedback ,
- Dataset Bias ,
- Sensitive Attributes ,
- Fair Model ,
- Section For Details ,
- Results In Table ,
- Sample Quality ,
- Latent Space ,
- High-quality Images ,
- Objective Quality ,
- Bigotry ,
- Section For More Details ,
- Generative Adversarial Networks Training ,
- Conventional Setup ,
- Black Hair ,
- Amount Of Bias ,
- Adaptation Stage ,
- Gram Matrix ,
- Simple Transfer ,
- Gaussian Noise Distribution ,
- Discriminative Learning ,
- Curly Hair
- Author Keywords
Keywords assist with retrieval of results and provide a means to discovering other relevant content. Learn more.
- IEEE Keywords
- Index Terms
- Transfer Learning ,
- Transfer Learning Approach ,
- Deep Generative Models ,
- Bias Mitigation ,
- Large Datasets ,
- Small Datasets ,
- Quality Performance ,
- Generative Adversarial Networks ,
- Quality Of Feedback ,
- Dataset Bias ,
- Sensitive Attributes ,
- Fair Model ,
- Section For Details ,
- Results In Table ,
- Sample Quality ,
- Latent Space ,
- High-quality Images ,
- Objective Quality ,
- Bigotry ,
- Section For More Details ,
- Generative Adversarial Networks Training ,
- Conventional Setup ,
- Black Hair ,
- Amount Of Bias ,
- Adaptation Stage ,
- Gram Matrix ,
- Simple Transfer ,
- Gaussian Noise Distribution ,
- Discriminative Learning ,
- Curly Hair
- Author Keywords