{"title":"Make your data fair: A survey of data preprocessing techniques that address biases in data towards fair AI","authors":"Amal Tawakuli, Thomas Engel","doi":"10.1016/j.jer.2024.06.016","DOIUrl":null,"url":null,"abstract":"<div><div>During the public trials of ChatGPT, it was highlighted that the language model can generate racially discriminatory responses. This issue, however is not new to AI. Several models and networks exhibited sexism, racism and other discriminatory traits in their output. Needless to say, discrimination and biases in AI must be addressed. The urgency of addressing this issue, however, is becoming more evident and pressing with the widespread adoption of AI solutions across different aspects of our lives. This paper is a gentle introduction of Fairness in AI and a survey of existing solutions. The root cause of unfair AI, is the data used to train and test the algorithms. As such, our survey focuses on data preprocessing techniques that address biases and discrimination in the data consumed by AI.</div></div>","PeriodicalId":48803,"journal":{"name":"Journal of Engineering Research","volume":"13 3","pages":"Pages 2362-2369"},"PeriodicalIF":2.2000,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Engineering Research","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2307187724001871","RegionNum":4,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2024/7/10 0:00:00","PubModel":"Epub","JCR":"Q3","JCRName":"ENGINEERING, MULTIDISCIPLINARY","Score":null,"Total":0}
引用次数: 0
Abstract
During the public trials of ChatGPT, it was highlighted that the language model can generate racially discriminatory responses. This issue, however is not new to AI. Several models and networks exhibited sexism, racism and other discriminatory traits in their output. Needless to say, discrimination and biases in AI must be addressed. The urgency of addressing this issue, however, is becoming more evident and pressing with the widespread adoption of AI solutions across different aspects of our lives. This paper is a gentle introduction of Fairness in AI and a survey of existing solutions. The root cause of unfair AI, is the data used to train and test the algorithms. As such, our survey focuses on data preprocessing techniques that address biases and discrimination in the data consumed by AI.
期刊介绍:
Journal of Engineering Research (JER) is a international, peer reviewed journal which publishes full length original research papers, reviews, case studies related to all areas of Engineering such as: Civil, Mechanical, Industrial, Electrical, Computer, Chemical, Petroleum, Aerospace, Architectural, Biomedical, Coastal, Environmental, Marine & Ocean, Metallurgical & Materials, software, Surveying, Systems and Manufacturing Engineering. In particular, JER focuses on innovative approaches and methods that contribute to solving the environmental and manufacturing problems, which exist primarily in the Arabian Gulf region and the Middle East countries. Kuwait University used to publish the Journal "Kuwait Journal of Science and Engineering" (ISSN: 1024-8684), which included Science and Engineering articles since 1974. In 2011 the decision was taken to split KJSE into two independent Journals - "Journal of Engineering Research "(JER) and "Kuwait Journal of Science" (KJS).