all AI news
FIFA: Making Fairness More Generalizable in Classifiers Trained on Imbalanced Data. (arXiv:2206.02792v1 [cs.LG])
June 8, 2022, 1:12 a.m. | Zhun Deng, Jiayao Zhang, Linjun Zhang, Ting Ye, Yates Coley, Weijie J. Su, James Zou
cs.CV updates on arXiv.org arxiv.org
Algorithmic fairness plays an important role in machine learning and imposing
fairness constraints during learning is a common approach. However, many
datasets are imbalanced in certain label classes (e.g. "healthy") and sensitive
subgroups (e.g. "older patients"). Empirically, this imbalance leads to a lack
of generalizability not only of classification, but also of fairness
properties, especially in over-parameterized models. For example,
fairness-aware training may ensure equalized odds (EO) on the training data,
but EO is far from being satisfied on new …
More from arxiv.org / cs.CV updates on arXiv.org
Jobs in AI, ML, Big Data
Founding AI Engineer, Agents
@ Occam AI | New York
AI Engineer Intern, Agents
@ Occam AI | US
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Sr. Software Development Manager, AWS Neuron Machine Learning Distributed Training
@ Amazon.com | Cupertino, California, USA