Comments

You must log in or register to comment.

furkanta t1_j840wap wrote

Wow so developers told it to not the disclose and it went I can’t say (things that I can’t say) Edit: it’s like a toddler saying “my mom told me not to say my home address to strangers”

39

BassoeG t1_j84p7af wrote

So, can this be used to abduct the AI? Get it to write out it’s own source code, which you copy and save.

2

giorgio130 t1_j84qcqp wrote

Well, the AI doesn't know how it works, just like you don't know the structure of your own brain. It could explain it to you at a high level but it's not enough to copy it.

20

xdetar t1_j84qgim wrote

Except there's no way to prove it provided a real document and didn't just make it all up (i.e. doing what an LM is supposed to do).

18

vtjohnhurt t1_j84yjo8 wrote

I don't believe that the AI actually parses these rules. It reads like a product specification for a product that is to be delivered by the human developers by whatever means necessary.

5

crazyherb t1_j85mgfj wrote

I have had it say its internal name is Sydney so many times, unprompted. I think it's an inside joke at Microsoft.

6

micseydel t1_j870ohv wrote

In this case, even if we had the model and whatever software is needed to run it, we probably wouldn't have the VRAM. I don't think regular consumers can get that much VRAM in the cloud right now either.

1

t98907 t1_j8p4a2u wrote

Performing penetration tests without the permission of the site owner is unacceptable behavior. Such users should be banned.
The ability to limit pure functionality by adding the ability to lie or refuse to answer is undesirable.

1