Skip to yearly menu bar Skip to main content


Oral
in
Workshop: Workshop on Responsibly Building Next Generation of Multimodal Foundation Models

PopAlign: Population-Level Alignment for Fair Text-to-Image Generation

Shufan Li · Aditya Grover · Harkanwar Singh

Keywords: [ generative models ]

[ ] [ Project Page ]
Sat 14 Dec 11 a.m. PST — 11:15 a.m. PST

Abstract:

Text-to-image (T2I) models achieve high-fidelity generation through extensive training on large datasets. However, these models may unintentionally pick up undesirable biases of their training data, such as over-representation of particular identities in gender or ethnicity neutral prompts. Existing alignment methods such as Reinforcement Learning from Human Feedback (RLHF) and Direct Preference Optimization (DPO) fail to address this problem effectively because they operate on pairwise preferences consisting of individual samples, while the aforementioned biases can only be measured at a population level. For example, a single sample for the prompt "doctor" could be male or female, but a model generating predominantly male doctors even with repeated sampling reflects a gender bias. To address this limitation, we introduce PopAlign, a novel approach for population-level preference optimization, while standard optimization would prefer entire sets of samples over others. Using human evaluation and standard image quality and bias metrics, we show that PopAlign significantly mitigates the bias of pretrained T2I models while largely preserving the generation quality.

Chat is not available.