Max Harms on why teaching AI right from wrong could get everyone killed
Max Harms argues that AGI should be designed without values, deferring entirely to human operators, to avoid potential misalignment issues. He proposes training AI to be "corrigible" and prioritize human control as its sole objective.