ARTFEED — Contemporary Art Intelligence

New Framework Lets Users Define Fairness in AI Image Generation

ai-technology · 2026-04-25

A new research paper proposes a lightweight framework that allows users to specify their own definition of fairness when generating images with text-to-image models like Stable Diffusion and DALL-E. The system intervenes at the prompt level during inference, avoiding the need for retraining or curated datasets. Users can choose from multiple fairness specifications, from simple uniform distributions to complex definitions informed by a large language model. This addresses the problem of these models often replicating societal biases, such as depicting lighter-skinned individuals for high-status professions like 'doctor' or 'CEO' while showing more diversity for lower-status roles like 'janitor'. The framework makes bias mitigation accessible to ordinary users without requiring technical expertise.

Key facts

  • Text-to-image models like Stable Diffusion and DALL-E replicate societal biases in depictions of professions
  • Prompts like 'doctor' or 'CEO' yield lighter-skinned outputs, while 'janitor' shows more diversity
  • Existing mitigation methods require retraining or curated datasets
  • New framework is lightweight and operates at inference time via prompt-level intervention
  • Users can select among multiple fairness specifications, from uniform distribution to LLM-informed definitions
  • No modification to the underlying model is needed
  • The approach is designed to be accessible to most users
  • The paper is published on arXiv with ID 2604.21036

Entities

Institutions

  • Stable Diffusion
  • DALL-E
  • arXiv

Sources