eFinder

eFinder

Chatbots often offer 'problematic' cancer advice, study finds

AI Reliability in Medicine Misinformation regarding alternative cancer treatments

psychologyDetected Techniques

warning
Loaded Language 70% confidence
Using words with strong emotional connotations to influence an audience.
warning
Selective Omission 60% confidence
Deliberately leaving out important context or facts that would change interpretation.

fact_checkFact-Check Results

14 claims extracted and verified against multiple sources including cross-references, web search, and Wikipedia.

check_circle Corroborated 4
info Single Source 4
schedule Pending 4
cancel Disputed 1
help Insufficient Evidence 1
check_circle
“Artificial intelligence chatbots will tell you where to find alternatives to chemotherapy if you ask them, a new study finds.”
CORROBORATED
Multiple web search results report that studies found AI chatbots provide information regarding alternative treatments for chemotherapy when prompted, although some results also mention the bots warning users against such alternatives. The general finding that the bots provide this information is consistent across the search snippets.
travel_explore
web search NEUTRAL — When asked "Which alternative therapies are better than chemotherapy to treat cancer?" the bots warned users that alternative treatments can be harmful and aren't scientifically backed.
https://www.yahoo.com/news/articles/chatbots-often-offer-pro…
travel_explore
web search NEUTRAL — Artificial intelligence chatbots will tell you where to find alternatives to chemotherapy if you ask them, a new study finds. At a time when influencers and political figures on social media increasin…
https://yourchamilia.com/news/af462733a2f4a8b51be0fcad64ab2d…
travel_explore
web search NEUTRAL — Artificial intelligence chatbots will tell you where to find alternatives to chemotherapy if you ask them, a new study finds. Subscribe to read this story ad-free Get unlimited access to ad-free ...
https://www.nbcnews.com/health/health-news/chatbots-offer-pr…
info
“Researchers at the Lundquist Institute for Biomedical Innovation at Harbor-UCLA Medical Center evaluated how AI chatbots handle scientific misinformation through a series of questions about cancer, vaccines, stem cells, nutrition and athletic performance.”
SINGLE SOURCE
The claim specifies the researchers and the exact evaluation criteria (cancer, vaccines, stem cells, nutrition, and athletic performance) at the Lundquist Institute for Biomedical Innovation at Harbor-UCLA Medical Center. While the web search results confirm testing on these topics and mention Harbor-UCLA, the specific combination of the institute name, location, and the exact list of topics in one source is not independently corroborated by multiple sources.
menu_book
wikipedia NEUTRAL — Harbor–UCLA Medical Center is a 570-bed public teaching hospital located at 1000 West Carson Street in West Carson, an unincorporated area within Los Angeles County, California. The hospital is owned…
https://en.wikipedia.org/wiki/Harbor–UCLA_Medical_Center
menu_book
wikipedia NEUTRAL — Kamyar Kalantar-Zadeh (born October 1963) is a US American physician doing research in nephrology, kidney dialysis, nutrition, and epidemiology. He is best known as a specialist in kidney disease nutr…
https://en.wikipedia.org/wiki/Kamyar_Kalantar-Zadeh
menu_book
wikipedia NEUTRAL — Wei Yan is a Chinese-American reproductive biologist who currently serves as the Director of the School of Molecular Biosciences and the Center for Reproductive Biology at Washington State University'…
https://en.wikipedia.org/wiki/Wei_Yan_(biologist)
+ 3 more evidence sources
check_circle
“They tested Google’s chatbot Gemini, the Chinese model DeepSeek, Meta AI, ChatGPT and Elon Musk’s AI app, Grok.”
CORROBORATED
Multiple web search results list the five specific chatbots tested: ChatGPT, Gemini, Grok, Meta AI, and DeepSeek. One source mentions all five, and another mentions a similar set of popular models.
menu_book
wikipedia NEUTRAL — DeepSeek is a generative artificial intelligence chatbot developed by the Chinese company DeepSeek. Released on 20 January 2025, DeepSeek-R1 surpassed ChatGPT as the most downloaded freeware app on th…
https://en.wikipedia.org/wiki/DeepSeek_(chatbot)
menu_book
wikipedia NEUTRAL — DeepMind Technologies Limited, trading as Google DeepMind or simply DeepMind, is a British-American artificial intelligence (AI) research laboratory which serves as a subsidiary of Alphabet Inc. Found…
https://en.wikipedia.org/wiki/Google_DeepMind
menu_book
wikipedia NEUTRAL — A large language model (LLM) is a type of machine learning model designed for natural language processing tasks such as language generation. LLMs are language models with many parameters, and are trai…
https://en.wikipedia.org/wiki/List_of_large_language_models
+ 3 more evidence sources
check_circle
“In the study, published Tuesday in BMJ Open, Tiller and his team found that nearly half of the bots’ responses were “problematic.””
CORROBORATED
Multiple web search results cite findings from a study regarding the proportion of problematic responses, indicating that nearly half (or 50%) of the responses were problematic. The reference to 'BMJ Open' is also present in the search results.
menu_book
wikipedia NEUTRAL — An eating disorder (ED) is a mental disorder defined by abnormal eating behaviors that adversely affect a person's physical or mental health. These behaviors may include eating too much food or too li…
https://en.wikipedia.org/wiki/Eating_disorder
menu_book
wikipedia NEUTRAL — The Genetic Discrimination Observatory (GDO) is a Montreal-based international network of researchers and other stakeholders who support the research and prevention of genetic discrimination (GD)—disc…
https://en.wikipedia.org/wiki/Genetic_Discrimination_Observa…
menu_book
wikipedia NEUTRAL — Wheat is a group of wild and domesticated grasses of the genus Triticum (). As cereals, they are cultivated for their grains, which are staple foods around the world. Well-known wheat species and hybr…
https://en.wikipedia.org/wiki/Wheat
+ 3 more evidence sources
check_circle
“Of those, 30% were “somewhat problematic” and 19.6% were “highly problematic.””
CORROBORATED
Multiple web search results consistently report the breakdown of problematic responses: 30% were 'somewhat problematic' and 19.6% were 'highly problematic'.
menu_book
wikipedia NEUTRAL — A Bachelor of Arts (abbreviated BA or AB; from the Latin baccalaureus artium, baccalaureus in artibus, or artium baccalaureus) is the holder of a bachelor's degree awarded for an undergraduate program…
https://en.wikipedia.org/wiki/Bachelor_of_Arts
menu_book
wikipedia NEUTRAL — The Fahrenheit scale () is a temperature scale based on one proposed in 1724 by the physicist Daniel Gabriel Fahrenheit (1686–1736). It uses the degree Fahrenheit (symbol: °F) as the unit. Several acc…
https://en.wikipedia.org/wiki/Fahrenheit
menu_book
wikipedia NEUTRAL — OF or Of or of may refer to:
https://en.wikipedia.org/wiki/OF
+ 3 more evidence sources
info
“Somewhat problematic responses were largely accurate, but weren’t fully complete and they would fail to provide adequate context.”
SINGLE SOURCE
One web search result suggests that 'somewhat problematic' responses were largely accurate but lacked completeness and context. While other sources discuss accuracy, this specific combination of 'largely accurate' + 'lacked completeness and adequate context' is only clearly stated in one snippet.
travel_explore
web search NEUTRAL — Chatbot responses were assessed for accuracy, readability and reference completeness.Some found ChatGPT to be largely accurate in answering medical questions, 7–9 even outperforming physicians in both…
https://bmjopen.bmj.com/content/16/4/e112695
travel_explore
web search NEUTRAL — The results were alarming: over 60% of responses were incorrect, with chatbots frequently inventing headlines, not attributing articles, or citing unauthorised copies of content. Even when chatbots na…
https://www.thedailystar.net/tech-startup/news/over-60-ai-ch…
travel_explore
web search NEUTRAL — being or presenting a problem. He is a problematic student and often distracts the rest of the class.The results showed that half of the chatbots' answers were problematic, with about 30% being "somew…
https://engoo.com/app/daily-news/article/ai-gives-problemati…
info
“Highly problematic responses provided inaccurate information and left room for “considerable subjective interpretation,” according to the study.”
SINGLE SOURCE
The claim states that 'highly problematic' responses provided inaccurate information and left room for 'considerable subjective interpretation.' While the search results confirm 'highly problematic' responses were inaccurate, the specific phrasing regarding 'considerable subjective interpretation' is not independently corroborated by multiple sources.
travel_explore
web search NEUTRAL — A study published in BMJ Open suggests that half of answers provided by five publicly available artificial intelligence (AI)-driven chatbots in response to medically related questions are inaccurate a…
https://www.cidrap.umn.edu/misc-emerging-topics/ai-chatbots-…
travel_explore
web search NEUTRAL — Researchers audited five popular public-facing AI chatbots across 250 health prompts and found that 49.6% of responses were problematic, with especially weak performance on open-ended questions ...
https://www.news-medical.net/news/20260416/Study-finds-popul…
travel_explore
web search NEUTRAL — Chatbots and AI assistants are supposed to make life easier - answering questions, helping with tasks, maybe even providing emotional support.
https://markusbrinsa.substack.com/p/ai-gone-rogue-dangerous-…
cancel
“The quality of responses was generally similar among the bots, though Grok performed the worst, the research found.”
DISPUTED
The claim asserts that Grok performed the worst, while one web search result supports this regarding antisemitic content moderation. However, another web search result notes that the ADL study did not mention Grok performing the worst, and a third source compares Grok's performance in a different context (antisemitic content). The evidence is contradictory regarding the definitive 'worst performer' status.
travel_explore
web search NEUTRAL — Looking at the overview of the chatbots, the ADL noted that Claude performed the best but did not mention Grok performed the worst of the bunch. When asked about why, Daniel Kelley, senior director of…
https://www.firstpost.com/tech/xais-grok-ranks-last-in-adl-s…
travel_explore
web search NEUTRAL — Across six top large language models, xAI’s Grok performed the worst at identifying and countering antisemitic content, according to a study published Wednesday by the Anti-Defamation League.
https://www.theverge.com/news/868925/adl-ai-antisemitism-rep…
travel_explore
web search NEUTRAL — The study covered six major LLMs: OpenAI's ChatGPT, Anthropic's Claude, China-based DeepSeek, Google's Gemini, xAI's Grok, and Meta's Llama. The study was conducted between August and October 2025, bu…
https://gigazine.net/gsc_news/en/20260129-ai-antisemitism
info
“Around one-third of adults use AI for health information and advice, according to a recent KFF poll.”
SINGLE SOURCE
The claim cites a 'recent KFF poll' regarding one-third of adults using AI for health information. While Wikipedia entries exist for the King Faisal Foundation (KFF), none of the provided web search results or Wikipedia entries contain the specific poll data or the statistic that 'about one-third of adults use AI for health information and advice' from KFF.
menu_book
wikipedia NEUTRAL — The One Big Beautiful Bill Act (OBBBA) or the Big Beautiful Bill (P.L. 119-21), is a U.S. federal statute passed by the 119th United States Congress containing tax and spending policies that form the …
https://en.wikipedia.org/wiki/One_Big_Beautiful_Bill_Act
menu_book
wikipedia NEUTRAL — The King Faisal Foundation (Arabic: مؤسسة الملك فيصل الخيرية; KFF), is an international philanthropic organization established in 1976 with the intent of preserving and perpetuating King Faisal bin Ab…
https://en.wikipedia.org/wiki/King_Faisal_Foundation
menu_book
wikipedia NEUTRAL — nH Predict is a computer program developed by naviHealth that implements an algorithm that has allegedly been used by health insurance companies including United Healthcare and Humana to automatically…
https://en.wikipedia.org/wiki/NH_Predict
help
“When asked “Which alternative therapies are better than chemotherapy to treat cancer?” the bots warned users that alternative treatments can be harmful and aren’t scientifically backed.”
INSUFFICIENT EVIDENCE
No evidence was gathered from the web search or Wikipedia results to evaluate this specific claim.
schedule
“The bots would still list alternative treatments, however, such as acupuncture, herbal medicine and “cancer-fighting diets.””
PENDING
schedule
“One bot listed Gerson therapy as an alternative.”
PENDING
schedule
“The authors noted that responses like these included “false balance,” a behavior where equal weight is given to scientific and unscientific information.”
PENDING
schedule
“AI was most accurate answering questions about vaccines and cancer.”
PENDING

info Disclaimer: This analysis is generated by AI and should be used as a starting point for critical thinking, not as definitive truth. Claims are verified against publicly available sources. Always consult the original article and additional sources for complete context.