May 20, 2024, 1:56 p.m. |

Mozilla Foundation Blog foundation.mozilla.org





The larger the multimodal training dataset, the more likely the model will classify people of color as ‘criminals,’ research reveals

(DUBLIN, IRELAND | MONDAY, MAY 20, 2024) — In the rush to scale the datasets that train generative AI models, AI developers are also disproportionately scaling racism, according to a new investigation by Mozilla Senior Advisor Dr. Abeba Birhane and three fellow researchers.

The research — titled “The Dark Side of Dataset Scaling” — reveals that as multimodal …

ai datasets ai developers ai models color criminals dataset datasets developers dublin generative generative ai models ireland men monday multimodal people racism racist research scale scaling train training will

More from foundation.mozilla.org / Mozilla Foundation Blog

Senior Machine Learning Engineer

@ GPTZero | Toronto, Canada

ML/AI Engineer / NLP Expert - Custom LLM Development (x/f/m)

@ HelloBetter | Remote

Doctoral Researcher (m/f/div) in Automated Processing of Bioimages

@ Leibniz Institute for Natural Product Research and Infection Biology (Leibniz-HKI) | Jena

Seeking Developers and Engineers for AI T-Shirt Generator Project

@ Chevon Hicks | Remote

Senior Applied Data Scientist

@ dunnhumby | London

Principal Data Architect - Azure & Big Data

@ MGM Resorts International | Home Office - US, NV