Dec. 31, 2023, 2:01 p.m. | Yunzhe Wang

Towards AI - Medium pub.towardsai.net

A technical explainer of the paper “Weak-to-strong generalization: eliciting strong capabilities with weak supervision”

Superintelligent AI systems will be extraordinarily powerful; humans could face catastrophic risks including even extinction if those systems are misaligned or misused. It is important for AI developers to have a plan for aligning superhuman models ahead of time — before they have the potential to cause irreparable harm. (Appendex G in the paper)
illustration by Midjourney

Weak-to-Strong Generalization: Eliciting Strong Capabilities With Weak Supervision

Human …

ai ai developers ai-safety ai systems capabilities chatgpt developers explainer extinction face humans openai paper risks superalignment superhuman superintelligent ai supervision systems technical trains weak ai will

Software Engineer for AI Training Data (School Specific)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Python)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Tier 2)

@ G2i Inc | Remote

Data Engineer

@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US