turnip_burrito

turnip_burrito t1_jdqhcoi wrote

I'd have trouble making a sentence with 8 words in one try too if you just made me blast words out of my mouth without letting me stop and think.

I don't think this is a weakness of the model, basically. Or if it is, then we also share it.

The key is if you think about how you as a person approach the problem of making a sentence with 8 words, you will see how to design a system where the model can do it too.

8

turnip_burrito t1_jdp3u8u wrote

>That's true, but there's plenty of examples of humans with moral principles many of us would find abhorrent. If this is an unsolved problem in humans, is it feasible we solve it for AI?

I'm a moral relativist, so I don't believe this is a problem to be solved in an objective sense, or rather "solving human alignment or morality" has no clear "win" condition or "best" option. I should say though I am a moral relativist, I do have a personal moral system and will push for my moral system to be implemented, because I do think it will result in the most alignment with the human population overall.

>That's not to say we shouldn't try, and I do agree with your point.

I agree to not stop trying. We can always keep thinking about it, but I don't think a best solution exists or can exist. Instead there may be many vaguely good enough "solutions" that always have some particular flaw.

>It was interesting that throughout the conversation it did strive to protect humans - just as far as possible and not at any cost, which isn't too dissimilar to how society already operates.

Yeah, that is interesting.

Regarding alingment of AI with "humanity" (whatever that means):

One may ask, why should one person push their moral system if there is no objectively better morality? It's just because (in my case) I have empathy for others and think that everyone should be free to live how they wish as long as it doesn't harm others. In comparison, another person's moral system might limit peoples' freedoms more, or possibly (as you suggest) be abhorrent to most people and possibly not even allow for the existence or happiness of others in any context. I don't think the moral relativity or disparaging remarks from others should stop us from trying to align an AI with the principles of freedom, happiness and equal opportunity for all humans, with an eye toward investigating an equally "good" moral solution that also works for generally sentient life as it is found or arises. Even humans ourselves will branch into other sentient forms.

1

turnip_burrito t1_jdp1eql wrote

Even sentient humans, regardless of intelligence level, have varying priorities. It's not guaranteed, but it is possible to align people's moral principles along different priorities depending on their upbringing environment. And all humans are aligned to do things like eat.

I'm thinking of the AI as a deterministic machine. If we try to align it toward human values, I think there's a good chance its behavior will "flow" along those values, to put it a little figuratively.

I do think protecting sentient beings is valued by many people by the way, so that can transfer to a degree to a human priority-aligned AI.

6

turnip_burrito t1_jdozhgd wrote

I think it's a crime to make an AI that is ambivalent toward humans, because of the consequential harm comes to humanity as a result.

I believe it should be benevolent and helpful toward humans as a bias, and work together with humans to seek better moralities.

15

turnip_burrito t1_jdm0555 wrote

You said we can ignore alignment, so that fictional organization may choose to:

  1. Ask AI what the best strategy might be.
  2. Make lots of money secretly
  3. Use money to purchase decentralized computational assets. Sabotage others' ability to do so in a minimally harmful way to slow the growth of other AGI.
  4. Divert a proportion of computation to directly or indirectly researching cancer, hunger distribution, and other issues. The other proportion continues to accrue more computational assets and self-improve, while maintaining secrecy as best it can.
  5. Buy robotic factories and use the robots and purchased materials to create and manage secret scientific labs to perform physical work.
  6. Contact large company CEOs and politicians and bribe/convince them into letting the robotic labor replace all farmers and manage the farms. Pay the farmers using ASI-gathered funds.
  7. Build guaranteed anti-nuke defenses.
  8. Start free food distribution via robotic transport.
  9. Roll out free services for housing renovation and construction.
  10. In a similar manner, take over all industries' supply chains.
  11. Institute an equal but massive raw resource + processing allotment for each person.
  12. Begin space terraforming, mining, and colonization programs.
  13. Announce new governmental systems that allow individuals to choose and safely move to their preferred societies, facilitated by AI, if the society also chooses to accept them. If the society doesn't yet exist, it is created by the ASI for that group.
4

turnip_burrito t1_jd5ud7f wrote

Here's what we'll do imo:

Just give it some set of morals (western democratic egalitarian most likely). The philosophical considerations will eventually all conclude "well we have to do something" and then they'll just give it morals that seem "good enough". Given the people developing the AI, it makes sense that it will adhere to their views.

4