The Hidden Power of Diabetes Drugs in Preserving Kidney Function
Type 2 diabetes is a prevalent metabolic disorder often associated with the progression of diabetic kidney disease (DKD), a leading ...
{"id":11912,"date":"2024-06-25T20:55:30","date_gmt":"2024-06-25T20:55:30","guid":{"rendered":"https:\/\/scienmag.com\/researchers-develop-new-training-technique-that-aims-to-make-ai-systems-less-socially-biased\/"},"modified":"2024-06-25T20:55:30","modified_gmt":"2024-06-25T20:55:30","slug":"researchers-develop-new-training-technique-that-aims-to-make-ai-systems-less-socially-biased","status":"publish","type":"post","link":"https:\/\/scienmag.com\/researchers-develop-new-training-technique-that-aims-to-make-ai-systems-less-socially-biased\/","title":{"rendered":"Researchers develop new training technique that aims to make AI systems less socially biased"},"content":{"rendered":"
CORVALLIS, Ore. \u2013 An Oregon State University doctoral student and researchers at Adobe have created a new, cost-effective training technique for artificial intelligence systems that aims to make them less socially biased.<\/p>\n
<\/p>\n
Credit: Johanna Carson, OSU College of Engineering<\/p>\n
<\/p>\n
CORVALLIS, Ore. \u2013 An Oregon State University doctoral student and researchers at Adobe have created a new, cost-effective training technique for artificial intelligence systems that aims to make them less socially biased.<\/p>\n
Eric Slyman of the OSU College of Engineering and the Adobe researchers call the novel method FairDeDup<\/a>, an abbreviation for fair deduplication. Deduplication means removing redundant information from the data used to train AI systems, which lowers the high computing costs of the training.<\/p>\n Datasets gleaned from the internet often contain biases present in society, the researchers said. When those biases are codified in trained AI models, they can serve to perpetuate unfair ideas and behavior.<\/p>\n By understanding how deduplication affects bias prevalence, it\u2019s possible to mitigate negative effects \u2013 such as an AI system automatically serving up only photos of white men if asked to show a picture of a CEO, doctor, etc. when the intended use case is to show diverse representations of people.<\/p>\n \u201cWe named it FairDeDup as a play on words for an earlier cost-effective method, SemDeDup, which we improved upon by incorporating fairness considerations,\u201d Slyman said. \u201cWhile prior work has shown that removing this redundant data can enable accurate AI training with fewer resources, we find that this process can also exacerbate the harmful social biases AI often learns.\u201d<\/p>\n Slyman<\/a> presented the FairDeDup algorithm last week in Seattle at the IEEE\/CVF\u00a0Conference on Computer Vision and Pattern Recognition<\/a>.<\/p>\n FairDeDup works by thinning the datasets of image captions collected from the web through a process known as pruning. Pruning refers to choosing a subset of the data that\u2019s representative of the whole dataset, and if done in a content-aware manner, pruning allows for informed decisions about which parts of the data stay and which go.<\/p>\n \u201cFairDeDup removes redundant data while incorporating controllable, human-defined dimensions of diversity to mitigate biases,\u201d Slyman said. \u201cOur approach enables AI training that is not only cost-effective and accurate but also more fair.\u201d<\/p>\n In addition to occupation, race and gender, other biases perpetuated during training can include those related to age, geography and culture.<\/p>\n \u201cBy addressing biases during dataset pruning, we can create AI systems that are more socially just,\u201d Slyman said. \u201cOur work doesn\u2019t force AI into following our own prescribed notion of fairness but rather creates a pathway to nudge AI to act fairly when contextualized within some settings and user bases in which it\u2019s deployed. We let people define what is fair in their setting instead of the internet or other large-scale datasets deciding that.\u201d<\/p>\n