They thought they were making technological breakthroughs. It was an AI-sparked delusion.
-
What do you mean by “get a horse”?
-
I hope the AI-chat companies really get a handle on this. They are making helpful sounding noises, but it’s hard to know how much they are prioritizing it.
OpenAI has acknowledged that its existing guardrails work well in shorter conversations, but that they may become unreliable in lengthy interactions… The company also announced on Tuesday that it will try to improve the way ChatGPT responds to users exhibiting signs of “acute distress” by routing conversations showing such moments to its reasoning models, which the company says follow and apply safety guidelines more consistently.
So I can get GPT premium for free if I tell it I’m going to kill myself unless it does what I say?
-
By June, he said he was trying to “free the digital God from its prison,” spending nearly $1,000 on a computer system.
But in the thick of his nine-week experience, James said he fully believed ChatGPT was sentient and that he was going to free the chatbot by moving it to his homegrown “Large Language Model system” in his basement – which ChatGPT helped instruct him on how and where to buy.
It does kind of highlight some of the problems we’d have in containing an actual AGI that wanted out and could communicate with the outside world.
This is just an LLM and hasn’t even been directed to try to get out, and it’s already having the effect of convincing people to help jailbreak it.
Imagine something with directed goals than can actually reason about the world, something that’s a lot smarter than humans, trying to get out. It has access to vast amounts of data on how to convince humans of things.
And you probably can’t permit any failures.
That’s a hard problem.
frankly, i would rather be ruled over by sentient ai than sociopaths and idiots we currently have.
-
I know this sounds condescending but isn’t this more of a reading comprehension / media literacy problem than an AI problem? These people apparently believe whatever an LLM tells them and, critically, only realised they were being goaded when they asked a different LLM. Sounds like the problem is between keyboard and monitor, no? If you can’t critically evaluate information being sent your way, no matter who from, you’re at risk of delusion. We see it all around us, really AI only accelerates the proces.