An example of the misalignment problem. Humans and AI both agreed on the stated purpose (generate a recipe), AI just had some deeper goals in mind as well.
I doubt it had nefarious intentions. My money is on the bot just being stupid.
If I ask you to create a drink using Windex and Clorox would you do any different? Do you have alignment problem too?
Yes, I know better, but ask a kid that and perhaps they’d do it. A LLM isn’t thinking though, it’s repeating training through probabilities. And btw, yes, humans can be misaligned with each other, having self goals underneath common ones. Humans think though…well, most of them.
Wow, people purposefully entered non edible ingredients and results are weird? Who could expect.
Gotta love how a spokesperson for the company expressed their disappointment that people are misusing the tool, vs. being disappointed in the company for letting the AI tool go live when it is clearly not ready for prime time
They shouldn’t have named their AI skynet