March 2, 2024, 5:58 a.m. | Stephanie Kirmer

Towards Data Science - Medium towardsdatascience.com

When LLMs give us outputs that reveal flaws in human society, can we choose to listen to what they tell us?

Photo by Vince Fleming on Unsplash

Machine Learning, Nudged

By now, I’m sure most of you have heard the news about Google’s new LLM*, Gemini, generating pictures of racially diverse people in Nazi uniforms. This little news blip reminded me of something that I’ve been meaning to discuss, which is when models have blind spots, so we apply …

ai bias in ai diverse editors pick flaws gemini google human llm llms machine learning people photo racially diverse society

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Principal Machine Learning Engineer (AI, NLP, LLM, Generative AI)

@ Palo Alto Networks | Santa Clara, CA, United States

Consultant Senior Data Engineer F/H

@ Devoteam | Nantes, France