turnip_burrito

turnip_burrito t1_jablzeb wrote

In addition, there's also a large risk of somebody accidentally making it evil. We should probably stop training on data that has these narratives in it.

We shouldn't be surprised when we train a model on X, Y, Z and it can do Z. I'm actually surprised that so many people are surprised at ChatGPT's tendency to reproduce (negative) patterns from its own training data.

The GPTs we've created are basically split personality disorder AI because of all the voices on the Internet we've crammed into the model. If we provide it a state (prompt) that pushes it to some area of its state space, then it will evolve according to whatever pattern that state belongs to.

tl;dr: It won't take an evil human to create evil AI. All it could take is some edgy 15 year old script kid messing around with publicly-available near-AGI.

1

turnip_burrito t1_jaavbn1 wrote

I wish OpenAI hadn't ever released ChatGPT. Also all those "democratize AI" guys screaming for research labs to release their AI for public use. What a mess. Now we're gonna end up with third party non-research corporations trying to use AI to make money to our long term detriment, and probably get royally rekt by some corpo money-making AI.

−3

turnip_burrito t1_ja6mf21 wrote

> style is just a desperate grasp that humans have something to offer that ai doesn't.

I think this is the correct assessment. But we may choose human media/products/services for other subjective reasons, like a "journey, not the destination" mindset or social connection.

6

turnip_burrito t1_ja5y3hb wrote

I agree with all of this, but just to be a bit over-pedantic on one bit:

> Models cant speak or hear when they want to Its just not part of their programming.

As you said it's not part of their programming, in today's models. In general though, it wouldn't be too difficult to construct a new model that judges at each timestep based on both external stimuli and internal hidden states when to speak/interrupt or listen intently. Actually at first glance such a thing sounds trivial.

1

turnip_burrito t1_ja55xpa wrote

Do you realize that algorithmically, it is much easier to test approaches on finite state games and later scale up to games with infinite states?

1

turnip_burrito t1_ja2ngmw wrote

That's good.

Maybe also in the future, for an extra layer of safety, when you can several LLMs together, you can use separate LLMs "judges". The judges can have memory refreshed every time you interact with the main one, and can screen the main LLM for unwanted behavior. They can do this by taking the main LLM's tentative output string as their own input, and use that to stop the main LLM from misbehaving.

6