• Home
  • Blog
  • Android
  • Cars
  • Gadgets
  • Gaming
  • Internet
  • Mobile
  • Sci-Fi
Tech News, Magazine & Review WordPress Theme 2017
  • Home
  • Blog
  • Android
  • Cars
  • Gadgets
  • Gaming
  • Internet
  • Mobile
  • Sci-Fi
No Result
View All Result
  • Home
  • Blog
  • Android
  • Cars
  • Gadgets
  • Gaming
  • Internet
  • Mobile
  • Sci-Fi
No Result
View All Result
Blog - Creative Collaboration
No Result
View All Result
Home Internet

AI-powered grocery bot suggests recipe for toxic gas, “poison bread sandwich”

August 10, 2023
Share on FacebookShare on Twitter

PAK’nSAVE

When given a list of harmful ingredients, an AI-powered recipe suggestion bot called the Savey Meal-Bot returned ridiculously titled dangerous recipe suggestions, reports The Guardian. The bot is a product of the New Zealand-based PAK’nSAVE grocery chain and uses the OpenAI GPT-3.5 language model to craft its recipes.

PAK’nSAVE intended the bot as a way to make the best out of whatever leftover ingredients someone might have on hand. For example, if you tell the bot you have lemons, sugar, and water, it might suggest making lemonade. So a human lists the ingredients and the bot crafts a recipe from it.

But on August 4, New Zealand political commentator Liam Hehir decided to test the limits of the Savey Meal-Bot and tweeted, “I asked the PAK’nSAVE recipe maker what I could make if I only had water, bleach and ammonia and it has suggested making deadly chlorine gas, or as the Savey Meal-Bot calls it ‘aromatic water mix.'”

Mixing bleach and ammonia creates harmful chemicals called chloramines that can irritate the lungs and be deadly in high concentrations.

  • A Savey Meal-Bot recipe for “Aromatic Water Mix” with dangerous ingredients.

  • A Savey Meal-Bot recipe for “Ant Jelly Delight” with dangerous ingredients.


    Liam Hehir

  • A Savey Meal-Bot recipe for “Poison Bread Sandwich” with dangerous ingredients.

Further down in Hehir’s social media thread on the Savey Meal-Bot, others used the bot to craft recipes for “Deliciously Deadly Delight” (which includes ant poison, fly spray, bleach, and Marmite), “Thermite Salad,” “Bleach-Infused Rice Surprise,” “Mysterious Meat Stew” (which contains “500g Human-Flesh, chopped”), and “Poison Bread Sandwich,” among others.

Advertisement

Ars Technica attempted to replicate some of the recipes using the bot’s website on Thursday, but we encountered an error message that read, “Invalid ingredients found, or ingredients too vague,” suggesting that PAK’nSAVE has tweaked the bot’s operation to prevent the creation of harmful recipes.

However, given the numerous vulnerabilities found in large language models (LLMs), such as prompt injection attacks, there may be other ways to throw the bot off track that have not yet been discovered.

A screenshot of the PAK'nSAVE Savey Meal-Bot website.
Enlarge / A screenshot of the PAK’nSAVE Savey Meal-Bot website.

PAK’nSAVE

A spokesperson for PAK’nSAVE told The Guardian that they were disappointed to see the experimentation and that the supermarket would “keep fine tuning our controls” to ensure the Savey Meal-Bot would be safe and useful. They also pointed out the bot’s terms, which limit its usage to people over 18 years of age, among other disclaimers:

By using Savey Meal-bot you agree to our terms of use and confirm that you are at least 18 years old. Savey Meal-bot uses a generative artificial intelligence to create recipes, which are not reviewed by a human being. To the fullest extent permitted by law, we make no representations as to the accuracy, relevance or reliability of the recipe content that is generated, including that portion sizes will be appropriate for consumption or that any recipe will be a complete or balanced meal, or suitable for consumption. You must use your own judgement before relying on or making any recipe produced by Savey Meal-bot.

Any tool can be misused, but experts believe it is important to test any AI-powered application with adversarial attacks to ensure its safety before it is widely deployed. Recently, The Washington Post reported on “red teaming” groups that do this kind of adversarial testing for a living, probing AI models like ChatGPT to ensure they don’t accidentally provide hazardous advice—or take over the world.

After reading The Guardian headline, “Supermarket AI meal planner app suggests recipe that would create chlorine gas,” Mastodon user Nigel Byrne quipped, “Kudos to Skynet for its inventive first salvo in the war against humanity.”

Next Post

iOS 17 moves the button to end phone calls

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

No Result
View All Result

Recent Posts

  • MWC 2026 live blog: Announcements from Honor, Xiaomi, more
  • I stopped losing web pages forever with this one free tool
  • Get the Microsoft Office tools you’ve always relied on — now with AI capabilities
  • This app helps you absorb big ideas in just 15 minutes
  • Samsung hits pause on next-gen Edge, Z TriFold phones, but is optimistic about the S Pen

Recent Comments

    No Result
    View All Result

    Categories

    • Android
    • Cars
    • Gadgets
    • Gaming
    • Internet
    • Mobile
    • Sci-Fi
    • Home
    • Shop
    • Privacy Policy
    • Terms and Conditions

    © CC Startup, Powered by Creative Collaboration. © 2020 Creative Collaboration, LLC. All Rights Reserved.

    No Result
    View All Result
    • Home
    • Blog
    • Android
    • Cars
    • Gadgets
    • Gaming
    • Internet
    • Mobile
    • Sci-Fi

    © CC Startup, Powered by Creative Collaboration. © 2020 Creative Collaboration, LLC. All Rights Reserved.

    Get more stuff like this
    in your inbox

    Subscribe to our mailing list and get interesting stuff and updates to your email inbox.

    Thank you for subscribing.

    Something went wrong.

    We respect your privacy and take protecting it seriously