April 24, 2023, 12:48 a.m. | Vithya Yogarajan, Gillian Dobbie, Henry Gouk

cs.CL updates on arXiv.org arxiv.org

An indigenous perspective on the effectiveness of debiasing techniques for
pre-trained language models (PLMs) is presented in this paper. The current
techniques used to measure and debias PLMs are skewed towards the US racial
biases and rely on pre-defined bias attributes (e.g. "black" vs "white"). Some
require large datasets and further pre-training. Such techniques are not
designed to capture the underrepresented indigenous populations in other
countries, such as M\=aori in New Zealand. Local knowledge and understanding
must be incorporated to …

algorithms analysis arxiv bias biases datasets knowledge language language models large datasets new zealand paper perspective pre-training society training unbiased understanding

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Data Scientist (Database Development)

@ Nasdaq | Bengaluru-Affluence