Measuring Gender Bias in Text-to-Image Models using Object Detection
A downloadable project
This work presents a novel strategy to measure bias in text-to-image models. Using paired prompts that specify a gender and vaguely reference an object (e.g. “a man/woman holding an item”) we can examine whether certain objects are associated with a certain gender. We found that male prompts were more prone to generate objects including ties, backpacks, knives, and trucks. Female prompts were more likely to generate objects including handbags, umbrellas, bottles, and cups. We go on to outline a simple framework for regulators looking to measure gender bias in text-to-image models.
Status | Released |
Category | Other |
Author | harvey.mannering |
Tags | Text based |
Download
Download
AI Gov Hackathon.pdf 653 kB
Leave a comment
Log in with itch.io to leave a comment.