March 2, 2024, 5:58 a.m. | Stephanie Kirmer

Towards Data Science - Medium towardsdatascience.com

When LLMs give us outputs that reveal flaws in human society, can we choose to listen to what they tell us?

Photo by Vince Fleming on Unsplash

Machine Learning, Nudged

By now, I’m sure most of you have heard the news about Google’s new LLM*, Gemini, generating pictures of racially diverse people in Nazi uniforms. This little news blip reminded me of something that I’ve been meaning to discuss, which is when models have blind spots, so we apply …

ai bias in ai diverse editors pick flaws gemini google human llm llms machine learning people photo racially diverse society

Software Engineer for AI Training Data (School Specific)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Python)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Tier 2)

@ G2i Inc | Remote

Data Engineer

@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US