Published October 28, 2025 | Version v1
Conference paper Open

Prompting Away Stereotypes? Evaluating Bias in Text-to-Image Models for Occupations

Authors/Creators

Description

Text-to-Image (TTI) models are powerful creative tools but risk amplifying harmful social biases. We frame representational societal bias assessment as an image curation and evaluation task and introduce a pilot benchmark of occupational portrayals spanning five socially salient roles (CEO, Nurse, Software Engineer, Teacher, Athlete). Using five state-of-the-art models: closed-source (DALLE 3, Gemini Imagen 4.0) and open-source (FLUX.1-dev, Stable Diffusion XL Turbo, Grok-2 Image), we compare neutral baseline prompts against fairness-aware controlled prompts designed to encourage demographic diversity. All outputs are annotated for gender (male, female) and race (Asian, Black, White), enabling structured distributional analysis. Results show that prompting can substantially shift demographic representations, but with highly model-specific effects: some systems diversify effectively, others overcorrect into unrealistic uniformity, and some show little responsiveness. These findings highlight both the promise and the limitations of prompting as a fairness intervention, underscoring the need for complementary model-level strategies. We release all code and data for transparency and reproducibility maximus-powers/img-gen-bias-analysis: Analysis results of racial and gender bias in image generation across occupations by FLUX Dev, dall-e 3, SDXL Turbo, Grok-2 Image, Gemini Imagen 4.0

Files

GenProCC_submission (2).pdf

Files (5.0 MB)

Name Size Download all
md5:207116dc5db7e2a0e7d9a6e8b3d733b3
5.0 MB Preview Download

Additional details

Funding

European Commission
AIXPERT - An agentic, multi-layer, GenAI-powered backbone to make an AI system explainable, accountable, and transparent 101214389