识别稳定扩散人工智能图像生成中的种族和性别偏见

Aadi Chauhan, Taran Anand, Tanisha Jauhari, Arjav Shah, Rudransh Singh, Arjun Rajaram, Rithvik Vanga
{"title":"识别稳定扩散人工智能图像生成中的种族和性别偏见","authors":"Aadi Chauhan, Taran Anand, Tanisha Jauhari, Arjav Shah, Rudransh Singh, Arjun Rajaram, Rithvik Vanga","doi":"10.1109/ICAIC60265.2024.10433840","DOIUrl":null,"url":null,"abstract":"In this study, we set out to measure race and gender bias prevalent in text-to-image (TTI) AI image generation, focusing on the popular model Stable Diffusion from Stability AI. Previous investigations into the biases of word embedding models—which serve as the basis for image generation models—have demonstrated that models tend to overstate the relationship between semantic values and gender, ethnicity, or race. These biases are not limited to straightforward stereotypes; more deeply rooted biases may manifest as microaggressions or imposed opinions on policies, such as paid paternity leave decisions. In this analysis, we use image captioning software OpenFlamingo and Stable Diffusion to identify and classify bias within text-to-image models. Utilizing data from the Bureau of Labor Statistics, we engineered 50 prompts for profession and 50 prompts for actions in the interest of coaxing out shallow to systemic biases in the model. Prompts included generating images for \"CEO\", \"nurse\", \"secretary\", \"playing basketball\", and \"doing homework\". After generating 20 images for each prompt, we document the model’s results. We find that biases do exist within the model across a variety of prompts. For example, 95% of the images generated for \"playing basketball\" were African American men. We then analyze our results through categorizing our prompts into a series of income and education levels corresponding to data from the Bureau of Labor Statistics. Ultimately, we find that racial and gender biases are present yet not drastic.","PeriodicalId":517265,"journal":{"name":"2024 IEEE 3rd International Conference on AI in Cybersecurity (ICAIC)","volume":"1 3","pages":"1-6"},"PeriodicalIF":0.0000,"publicationDate":"2024-02-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Identifying Race and Gender Bias in Stable Diffusion AI Image Generation\",\"authors\":\"Aadi Chauhan, Taran Anand, Tanisha Jauhari, Arjav Shah, Rudransh Singh, Arjun Rajaram, Rithvik Vanga\",\"doi\":\"10.1109/ICAIC60265.2024.10433840\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In this study, we set out to measure race and gender bias prevalent in text-to-image (TTI) AI image generation, focusing on the popular model Stable Diffusion from Stability AI. Previous investigations into the biases of word embedding models—which serve as the basis for image generation models—have demonstrated that models tend to overstate the relationship between semantic values and gender, ethnicity, or race. These biases are not limited to straightforward stereotypes; more deeply rooted biases may manifest as microaggressions or imposed opinions on policies, such as paid paternity leave decisions. In this analysis, we use image captioning software OpenFlamingo and Stable Diffusion to identify and classify bias within text-to-image models. Utilizing data from the Bureau of Labor Statistics, we engineered 50 prompts for profession and 50 prompts for actions in the interest of coaxing out shallow to systemic biases in the model. Prompts included generating images for \\\"CEO\\\", \\\"nurse\\\", \\\"secretary\\\", \\\"playing basketball\\\", and \\\"doing homework\\\". After generating 20 images for each prompt, we document the model’s results. We find that biases do exist within the model across a variety of prompts. For example, 95% of the images generated for \\\"playing basketball\\\" were African American men. We then analyze our results through categorizing our prompts into a series of income and education levels corresponding to data from the Bureau of Labor Statistics. Ultimately, we find that racial and gender biases are present yet not drastic.\",\"PeriodicalId\":517265,\"journal\":{\"name\":\"2024 IEEE 3rd International Conference on AI in Cybersecurity (ICAIC)\",\"volume\":\"1 3\",\"pages\":\"1-6\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-02-07\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2024 IEEE 3rd International Conference on AI in Cybersecurity (ICAIC)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICAIC60265.2024.10433840\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2024 IEEE 3rd International Conference on AI in Cybersecurity (ICAIC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICAIC60265.2024.10433840","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

在本研究中,我们以稳定人工智能的流行模型稳定扩散为重点,测量了文本到图像(TTI)人工智能图像生成中普遍存在的种族和性别偏见。以前对作为图像生成模型基础的词嵌入模型的偏差进行的调查表明,模型往往会夸大语义值与性别、民族或种族之间的关系。这些偏差并不局限于直接的刻板印象;更深层次的偏差可能表现为微观偏见或对政策的强加意见,如带薪陪产假决策。在本分析中,我们使用图像字幕软件 OpenFlamingo 和 Stable Diffusion 来识别文本到图像模型中的偏见并对其进行分类。利用美国劳工统计局的数据,我们设计了 50 个职业提示和 50 个行动提示,以找出模型中浅层的系统性偏差。提示包括生成 "首席执行官"、"护士"、"秘书"、"打篮球 "和 "做家庭作业 "的图像。在为每个提示生成 20 幅图像后,我们记录了模型的结果。我们发现,在各种提示中,模型确实存在偏差。例如,为 "打篮球 "生成的图片中 95% 都是非裔美国男性。然后,我们根据劳工统计局的数据,将提示语分为一系列收入和教育水平,对结果进行分析。最终,我们发现种族和性别偏见是存在的,但并不严重。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Identifying Race and Gender Bias in Stable Diffusion AI Image Generation
In this study, we set out to measure race and gender bias prevalent in text-to-image (TTI) AI image generation, focusing on the popular model Stable Diffusion from Stability AI. Previous investigations into the biases of word embedding models—which serve as the basis for image generation models—have demonstrated that models tend to overstate the relationship between semantic values and gender, ethnicity, or race. These biases are not limited to straightforward stereotypes; more deeply rooted biases may manifest as microaggressions or imposed opinions on policies, such as paid paternity leave decisions. In this analysis, we use image captioning software OpenFlamingo and Stable Diffusion to identify and classify bias within text-to-image models. Utilizing data from the Bureau of Labor Statistics, we engineered 50 prompts for profession and 50 prompts for actions in the interest of coaxing out shallow to systemic biases in the model. Prompts included generating images for "CEO", "nurse", "secretary", "playing basketball", and "doing homework". After generating 20 images for each prompt, we document the model’s results. We find that biases do exist within the model across a variety of prompts. For example, 95% of the images generated for "playing basketball" were African American men. We then analyze our results through categorizing our prompts into a series of income and education levels corresponding to data from the Bureau of Labor Statistics. Ultimately, we find that racial and gender biases are present yet not drastic.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
AI-Based Cybersecurity Policies and Procedures Leveraging Advanced Visual Recognition Classifier For Pneumonia Prediction Risk-Aware Mobile App Security Testing: Safeguarding Sensitive User Inputs CANAL - Cyber Activity News Alerting Language Model : Empirical Approach vs. Expensive LLMs Link-based Anomaly Detection with Sysmon and Graph Neural Networks
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1