Automated AI engineers are not enough to solve AGI alignment.
Do we know if the most effective 'alignment solution' for ASIs is to convince us they're aligned while pursuing their own objectives?
Do we know if the most effective 'alignment solution' for ASIs is to convince us they're aligned while pursuing their own objectives?