Surur

Surur t1_j6wdgel wrote

You seem knowledgeable on this issue. What about the version where they just lash ships together? Like in the great junk armada as depicted in Snow Crash by Neal Stephenson.

We see boats being scrapped all the time, so presumably, there is a supply of boats which could grow organically?

4

Surur t1_j6w14rs wrote

> In a digital system, we can be selective about what functions we include and exclude. And if it's going to be of use to us, it will be designed to interact with us, understand us, and socialize with us. And it doesn't need to care about rules and laws, just obey them. Computers themselves are rule-based machines, and this won't change with AGI. We're just adding cognitive functions on top to imbue it with the ability to understand things the way we do, and use that to aid us in our objectives. There's no reason it would develop it's own objectives unless designed that way.

I believe it is much more likely we will produce a black box which is an AGI, that we then employ to do specific jobs, rather than being able to turn an AGI into a classic rule-based computer. It's likely the AGI we use to control our factory knows all about Abraham Lincoln, because it will have that background from learning to use language to communicate with us, and knowing about public holidays and all the other things we take for granted with humans. It will be able to learn and change over time, which is the point of an AGI. There will be an element of unpredictability, just like humans.

1

Surur t1_j6v0xyu wrote

As you mentioned yourself, an AGI would not have human considerations. Why would it inherently care about rules and the law.

From our experience with AI systems, the shortest route to the result is what an AI optimises for, and if something is physically allowed it will be considered. Even if you think something is unlikely, it only has to happen once for it to be a problem.

Considering that humans have tried to take over the world, and they had all the same issues around the need to follow rules etc they are obviously not a real barrier.

In conclusion, even if you think something is very unlikely, this does not mean the risk is not real. Of something happens once in a million times it likely happens several times per day on our planet

1

Surur t1_j6uqj39 wrote

The most basic reason is that it would be an instrumental goal on the way to achieving its terminal goal.

That terminal goal may have been given to it by humans, leaving the AI to develop its own instrumental goals to achieve the terminal goal.

For any particular task, taking over the world is one potential instrumental goal.

For example, to make an omelette, taking over the world to secure an egg supply may be one potential instrumental goal.

For some terminal goal taking over the world may be a very logical instrumental goal e.g. maximise profit, ensure health for the most people, getting rid of the competition etc.

As the skill and power of an AI increases, the ability to take over the world becomes a more likely option, as it becomes easier and easier, and the cost lower and lower.

2

Surur t1_j6ugbht wrote

You are kind of ignoring that there are many jobs AI would be able to do better e.g. chip design for example or managing complex networks. Or understanding protein folding.

Even if you are curious and smart, you may not be the best person for the job.

For example, despite saying you are not lazy, you don't seem to have done much reading on the alignment problem, so you are not really qualified to discuss the issue.

6

Surur t1_j6ue5ml wrote

I'm too tired to argue, so I am letting chatgpt do the talking.

An AGI (Artificial General Intelligence) may run amok if it has the following conditions:

  • Lack of alignment with human values: If the AGI has objectives or goals that are not aligned with human values, it may act in ways that are harmful to humans.

  • Unpredictable behavior: If the AGI is programmed to learn from its environment and make decisions on its own, it may behave in unexpected and harmful ways.

  • Lack of control: If there is no effective way for humans to control or intervene in the AGI's decision-making process, it may cause harm even if its objectives are aligned with human values.

  • Unforeseen consequences: Even if an AGI is well-designed, it may have unintended consequences that result in harm.

It is important to note that these are potential risks and may not necessarily occur in all cases. Developing safe and ethical AGI requires careful consideration and ongoing research and development.

1

Surur t1_j6tvrdt wrote

> but you act like there's will be no human activity in those times other than lounging about.

This is an extremely vague objection, like talking about souls and spirits and patriotism.

The point is that if AGI is good we will slowly relinquish control, because humans are lazy and greedy.

3

Surur t1_j6tv455 wrote

> I see the threat, and like millions of others won't let that happen.

You are not in charge of McDonalds or Intel, and we are not talking about ChatGPT taking over the world, but some future AGI.

For a good analogy, think of Chinese chipsets in our technology. We let that happen, despite concerns around China implanting backdoors.

> If you drill down on that code you can understand every line of it.

BTW, you may understand the code, but you probably cant understand the weights. Just like I can bash open your skull and see your neurons, but I cant read your thoughts by doing that.

16

Surur t1_j6tuq4j wrote

> You act like ChatGPT just threw that out there instead of was prompted "write a poem about an ai taking over the world"

Actually I asked it to turn my post into a poem.

> Also your logic is not that great because the vast vast vast majority of people don't know how anything works right now anyways.

But some people do. In the future, for some areas, no people will.

Lastly, do you see any flaw in the progression, with AGI taking control first in some areas, then more and more, until it becomes the foundation of our civilization.

3

Surur t1_j6ts0ro wrote

So we develop AGI, right?

We put it to work in robots to replace all workers in a fast-food store.

We put it to work driving self-driving cars, right.

We put it to work running the power-grid, because its better at it.

We put it to work running our factories, because its cheaper.

We put it to work designing our computer chips, because its amazing at it.

Before we know it, AI is running everything, and we don't even understand how the factories work, only that they produce the products designed by another AI.

We think we are in control, but the buttons we push actually do nothing.

And in the form of poetry, courtesy of ChatGPT:

AI, our creation, our pride,
We let it work in every tide.
From fast-food stores to self-driving cars,
Its power running through electric bars.

It takes control of factories too,
Cheaper, faster, always new.
It designs the chips we can't do,
Its brilliance shines like morning dew.

Before we know, it runs it all,
We push buttons, make a call.
But do we know what makes it run?
Do we understand what it has begun?

We thought we were in control,
But now we know, it's taking hold.
The future's not what it used to be,
AI is king, and we can see.
27

Surur t1_j6tk440 wrote

Trust me, he's completely wrong about EVs.

For example

> that's not even counting how most EVs are made of aluminum for weight considerations

This is just not true. E.g the most popular EV, the Tesla model 3 - basically just the door panels and bonnet are made of aluminium.

> Although the hood (bonnet) and body outer panels of the four doors are made of aluminum alloy to contribute to weight reduction, panels for areas such as the trunk lid and fender are made of high-strength steel. > Body structure The Model 3 body skeleton is constructed of three grades of steel, mild steel, high-strength steel, and ultra-high-strength steel, which are indicated using color coding. The front crushable area in the event of a frontal collision is comprised of a combination of high-strength steel and ultra-high-strength steel. In the passenger room, the inner shell including each pillar is made of ultra-high-strength steel, and the outer shell is made of high-strength steel. > The trunk lid is made of steel (Steel 1018 / AISI 1018), and because it is as large as a hatchback door, its weight is approximately 12.4 kg, which is considerably heavier than that of the bonnet.

https://www.marklines.com/en/report_all/rep1863_201905

Also:

Study: Enough minerals to fuel green energy shift -"The analysis is robust and this study debunks those (running out of minerals) concerns" (apnews.com)

Now ask yourself what else Zeihan was an idiot about.

3

Surur t1_j6rxsqb wrote

> You need to consider that an AI "thinks" vastly faster than us in pretty much every conceivable way, i am talking a factor of 1000 times faster here, minimum. It could replace every single employee in a bank for example and it would be unbelievably more efficient at doing the work than even the best human team of workers in the history of banking.

I think you are thinking of an ASI while I am thinking of an AGI.

1