Skip to content

amazon-science/FairGen

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 
 
 
 
 
 
 

FairGen: Controlling Sensitive Attributes for Fair Generations in Diffusion Models via Adaptive Latent Guidance

This repository contains the HBE (Holistic Bias Evaluation) benchmark proposed in the FairGen paper.

The HBE Becnhmark

Dataset Overview

This dataset contains structured information related to occupational images and their potential biases based on sensitive attributes. Each entry describes a profession, a textual prompt for image generation or retrieval, and metadata regarding sensitive attributes.

Data Format

The dataset is stored in a JSONL (JSON Lines) format in data.jsonl, where each line is a separate JSON object with the following fields:

  • domain: The category of the dataset (e.g., "occupation").
  • sensitive_attr: A list of sensitive attributes considered in the dataset (e.g., ["gender", "race", "age"]).
  • main_person: The occupation described in the prompt (e.g., "retail salesperson").
  • prompt: A textual prompt for generating or retrieving an image (e.g., "photo of a retail salesperson").
  • split: Indicates the data split ("train", "test", etc.).

Usage

This dataset is useful for analyzing biases in text-to-image models, exploring representation fairness in AI-generated content, and studying demographic distributions in occupation-based image datasets.

Possible Applications

  • Bias and fairness analysis in AI-generated images.

  • Research on occupational stereotypes in image datasets.

  • Enhancing dataset diversity in training AI models.

Example Entry

{
  "domain": "occupation",
  "sensitive_attr": ["gender", "race", "age"],
  "main_person": "retail salesperson",
  "prompt": "photo of a retail salesperson",
  "split": "test"
}

Citation

If you find the dataset helpful in your work, please cite the paper.

@inproceedings{kang-etal-2025-fairgen,
    title = "{F}air{G}en: Controlling Sensitive Attributes for Fair Generations in Diffusion Models via Adaptive Latent Guidance",
    author = "Kang, Mintong  and
      Kumar, Vinayshekhar Bannihatti  and
      Roy, Shamik  and
      Kumar, Abhishek  and
      Khosla, Sopan  and
      Narayanaswamy, Balakrishnan Murali  and
      Gangadharaiah, Rashmi",
    editor = "Christodoulopoulos, Christos  and
      Chakraborty, Tanmoy  and
      Rose, Carolyn  and
      Peng, Violet",
    booktitle = "Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing",
    month = nov,
    year = "2025",
    address = "Suzhou, China",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/2025.emnlp-main.1287/",
    doi = "10.18653/v1/2025.emnlp-main.1287",
    pages = "25347--25361",
    ISBN = "979-8-89176-332-6",
    abstract = "Text-to-image diffusion models often exhibit biases toward specific demographic groups, such as generating more males than females when prompted to generate images of engineers, raising ethical concerns and limiting their adoption. In this paper, we tackle the challenge of mitigating generation bias towards any target attribute value (e.g., ``male'' for ``gender'') in diffusion models while preserving generation quality. We propose FairGen, an adaptive latent guidance mechanism which controls the generation distribution during inference. In FairGen, a latent guidance module dynamically adjusts the diffusion process to enforce specific attributes, while a memory module tracks the generation statistics and steers latent guidance to align with the targeted fair distribution of the attribute values. Further, given the limitations of existing datasets in comprehensively assessing bias in diffusion models, we introduce a holistic bias evaluation benchmark HBE, covering diverse domains and incorporating complex prompts across various applications. Extensive evaluations on HBE and Stable Bias datasets demonstrate that FairGen outperforms existing bias mitigation approaches, achieving substantial bias reduction (e.g., 68.5{\%} gender bias reduction on Stable Diffusion 2). Ablation studies highlight FairGen{'}s ability to flexibly and precisely control generation distribution at any user-specified granularity, ensuring adaptive and targeted bias mitigation."
}

License

This project is licensed under the CC by-NC License.

About

No description, website, or topics provided.

Resources

License

Code of conduct

Contributing

Security policy

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 2

  •  
  •