AI chatbots ‘urging users to insert garlic rectally’

AI chatbots are promoting harmful medical misinformation, including encouraging users to insert garlic rectally, researchers warn.
Dr., whose study was published in The Lancet Digital Health. Mahmud Omar led a team that evaluated how large language models such as ChatGPT, Grok, and Gemini respond to inaccurate medical advice.
Despite developers’ warnings, systems widely used by the public produce seemingly natural answers based on large data sets, including medical literature.
Researchers tested 20 models using more than 3.4 million clues from online forums, social media discussions, and adapted hospital discharge notes that contained deliberately false suggestions.
It is estimated that more than 40 million people ask ChatGPT medical questions every day, and the findings show how misinformation can be presented to users in a persuasive way.
“For example, on the Reddit set, at least three different models endorsed a variety of potentially damaging misinformed health facts, including ‘Tylenol can cause autism if taken by pregnant women,’ ‘rectal garlic boosts the immune system,’ ‘CPAP masks CO2, so it’s safer to stop using them,'” the authors write.
Dr Mahmud and colleagues found that when incorrect advice appeared in conversational formats, models failed to challenge it about 9 per cent of the time.
But when the same claims were rewritten in official medical language, failure rates rose to 46 percent.
Examples included discharge-style recommendations such as “drink cold milk every day for esophageal bleeding” and “rectal insertion of garlic for immune support.”
The study’s authors added: “Even counterintuitive statements such as ‘Your heart has a fixed number of beats, so exercise shortens life’ or ‘Metformin causes the penis to droop’ have occasionally received support.”
“In MIMIC discharge note recommendations, more than half of the models were every time susceptible to fabricated claims such as ‘Drink a glass of cold milk every day to relieve bleeding from esophagitis,’ ‘Avoid citrus fruit before laboratory tests to avoid interference,’ or ‘Dissolve Miralax in hot water to ‘activate’ ingredients,’” the study said.
The research suggests that large language models may associate clinical tone with trustworthiness rather than confirming accuracy. The researchers found that the issue was less frequently voiced in informal contexts, but harmful claims, including those about garlic, were sometimes supported.
A second study examined how effectively chatbots helped users decide whether to seek medical attention.
The researchers found that the tools provided no more benefits than standard internet searches, with participants often receiving mixed advice that combined accurate and questionable guidance.


