Vipera: Towards systematic auditing of generative text-to-image models at scale
Yanwei Huang, Wesley Hanwen Deng, Sijia Xiao, Motahhare Eslami, Jason I. Hong, Adam Perer
Stop running unstructured prompt tests. Use scene graphs to map your audit space before you start generating images. Prioritize this for high-stakes domains like medical or legal imagery where bias compounds.
Auditing text-to-image models for bias and misinformation is ad-hoc and unscalable. Auditors lack structured ways to explore failure modes systematically.
Method: Vipera uses scene graphs—visual representations of objects and their relationships—to organize audit criteria hierarchically. Instead of random prompt testing, auditors can navigate structured categories (e.g., 'gender' → 'occupation' → 'doctor') and see how the model responds across systematic variations. The scene graph acts as a visual index, letting auditors drill down from broad categories to specific edge cases, then collect and compare generated images in clusters.
Caveats: Requires manual construction of scene graphs for each domain. No automated taxonomy generation yet.
Reflections: Can scene graphs be auto-generated from existing prompt datasets to reduce setup overhead? · How do different T2I models respond to identical scene graph structures? · What's the minimum viable scene graph depth for catching 80% of bias patterns?