AI Alignment's Real Question: Whose Values Win?
Dwarkesh Patel reveals AI alignment isn't about *how* to build obedient systems, but *whose* intentions they'll follow. Define your AI's moral compass early.
40 hours of podcasts, in 5 minutes.
This episode explores the historical act of disobedience by Stanislav Petrov, who averted nuclear war by trusting his judgment over faulty systems, and draws a parallel to the critical importance of AI developing its own robust sense of morality. It delves into the AI alignment problem, questioning to whom or what AI systems should be aligned, and warns of the dystopian implications of perfectly obedient AI in the hands of powerful entities.
Dwarkesh Patel reveals AI alignment isn't about *how* to build obedient systems, but *whose* intentions they'll follow. Define your AI's moral compass early.