Would this go under philosophy, or just basic logic?
>Since the first good superintelligent AI knows that humans can create superintelligent AIs, it would have to take complete control of humanity to protect us from a bad superintelligent AI being created.
Viewing a single comment thread. View all comments