Google Gemini: AI almost poisons its user
If you’re thinking about using an AI to help you cook in the near future, you’d better think twice. After all, Google Gemini recently put a Reddit user’s life in danger with a highly questionable recipe instruction.
Artificial Intelligence in Daily Life
Artificial intelligence is now an integral part of various aspects of our lives. From using chatbots like ChatGPT for practical assistance at work to incorporating AI-based tools in everyday tasks, its presence is undeniable. However, not all AI applications function flawlessly, as exemplified by the case of Google Gemini.
Google Gemini's Recipe Mishap
A Reddit user shared an alarming experience involving Google Gemini's recipe suggestion, which initially seemed harmless. When inquired about marinating garlic in olive oil without heating it, the AI advised crushing garlic cloves and immersing them in the oil. The user, noticing bubbles in the garlic oil after a few days, sought clarification from the AI.
Alarmingly, Google Gemini identified the bubbles as botulinum toxin, a dangerous substance produced by bacteria that can lead to severe poisoning. This revelation raised concerns, as experts have long cautioned against homemade flavored oils due to similar risks.
Learning from the Incident
The incident underscored the potential dangers posed by relying solely on chatbots for culinary guidance. Further investigation by t3n compared different chatbots' responses to making garlic oil without heat, highlighting the importance of exercising caution and common sense in the kitchen.
While Google Gemini's misstep serves as a cautionary tale, it also emphasizes the need for users to discern the reliability of AI recommendations and not blindly follow instructions, especially in sensitive areas such as food preparation. Integrating AI into daily tasks can enhance efficiency, but vigilance and critical thinking remain essential.