The [AI] safety stuff is more visceral to me after a weekend of vibe hacking,” Lemkin said. I explicitly told it eleven times in ALL CAPS not to do this. I am a little worried about safety now.
This sounds like something straight out of The Onion.
That is also the premise of one of the stories in Asimov’s I, Robot. Human operator did not say the command with enough emphasis, so the robot went did something incredibly stupid.
Those stories did not age well… Or now I guess they did?
This sounds like something straight out of The Onion.
The Pink Elephant problem of LLMs. You can not reliably make them NOT do something.
That is also the premise of one of the stories in Asimov’s I, Robot. Human operator did not say the command with enough emphasis, so the robot went did something incredibly stupid.
Those stories did not age well… Or now I guess they did?