The [AI] safety stuff is more visceral to me after a weekend of vibe hacking,” Lemkin said. I explicitly told it eleven times in ALL CAPS not to do this. I am a little worried about safety now.
This sounds like something straight out of The Onion.
That is also the premise of one of the stories in Asimov’s I, Robot. Human operator did not say the command with enough emphasis, so the robot went did something incredibly stupid.
Those stories did not age well… Or now I guess they did?
This sounds like something straight out of The Onion.
That is also the premise of one of the stories in Asimov’s I, Robot. Human operator did not say the command with enough emphasis, so the robot went did something incredibly stupid.
Those stories did not age well… Or now I guess they did?
The Pink Elephant problem of LLMs. You can not reliably make them NOT do something.
Just say 12 times next time
Even after he used “ALL CAPS”?!? Impossible!