Submitted by derstarkerwille t3_10qxou0 in philosophy
bildramer t1_j6te87v wrote
It's easy to misspecify or misgeneralize our needs and wants. When we make AIs that do have drives (usually in toy universes where we research reinforcement learning or meta-learning, or artificial evolution), we often see a concerning combination: superhuman performance, and strong pursuit/maximization of the wrong goal. Here's a paper listing evolutionary examples. There's another list of pure RL examples but I don't have the link handy.
Viewing a single comment thread. View all comments