80,000 Hours Podcast

Max Harms on why teaching AI right from wrong could get everyone killed

Max Harms argues that AGI should be designed without values, deferring entirely to human operators, to avoid potential misalignment issues. He proposes training AI to be "corrigible" and prioritize human control as its sole objective.

Listen