Pak ‘n’ Save’s Savey Meal-bot cheerfully created unappealing recipes when customers experimented with non-grocery household items
A spokesperson for the supermarket said they were disappointed to see “a small minority have tried to use the tool inappropriately and not for its intended purpose”
Disappointed. Bloody hell. A company that’s too cheap to curate some actual recipes is “disappointed” because the tech they built is creating lethal recipes. I wonder what their liability is if someone tried and got hurt/sick.
Let’s turn the outrage dial back a bit. People had the bot make recipies with non-grocery items for laughs ( ie bleach and ammonia) and unsurprisingly the bot combined them.
The fact that it’s capable of doing so is obscenely dangerous and should draw serious legal attention.
Corporations acting all parentally “disappointed” because they were too cheap to properly code it isn’t okay.
if (itemCategory != ‘grocery’) { warn(“don’t eat this”)}
They couldn’t be bothered to put in the most basic of filters for an input like this. The company doesn’t get a pass for this level of incompetence
Such poor coding, how hard could it be to not let people use products that aren’t food? This along with their comment just makes them seem so lazy.
so this is how ai starts to try to genocide us …
A spokesperson for the supermarket said they were disappointed to see “a small minority have tried to use the tool inappropriately and not for its intended purpose”.
You should be thrilled that people are highlighting the issue before your batshit insane implementation literally kills someone.
How fucking hard is it to define stuff as “food” or “not food”?
To the AIs defense, that’s a damn effective depression meal.