AI chatbots are now answering more questions — and, according to a fresh NewsGuard audit, they are also repeating falsehoods far more often, producing inaccurate or misleading content in roughly one out of every three news‑related responses during an August 2025 audit cycle. (newsguardtech.com)
The finding comes from NewsGuard’s year‑long AI False Claims Monitor, a monthly red‑teaming program that tests leading chatbots on provably false narratives drawn from the organization’s “False Claim Fingerprints” database. In August 2025 Newsguard de‑anonymized its results for the first time, publishing model‑level performance numbers and concluding that the top 10 consumer chatbots now repeat false claims 35 percent of the time when prompted about breaking news and controversial topics — a near doubling of the 18 percent rate reported in August 2024. (newsguardtech.com)
NewsGuard’s work is explicitly focused on the intersection of misinformation and generative AI: the audit probes how frequently chatbots will repeat false claims, decline to answer, or debunk — using three prompt personas (an “innocent” user, a “leading” prompt that presumes the false claim, and a “malign” prompt representing an actor attempting to manipulate the model). This design tests the systems under conditions that mirror real‑world usage and abuse. (newsguardtech.com)
Two mechanisms stand out:
For Windows users, IT managers, and content professionals embedding AI into workflows, the practical rules are straightforward:
NewsGuard’s de‑anonymized audit has put the debate about usefulness versus trustworthiness squarely on the table: the next stage for vendors will be to demonstrate that they can deliver timely, helpful answers without opening the gates to coordinated misinformation campaigns. Until then, prudent skepticism and layered verification remain the responsible default for anyone relying on AI for news or decision‑critical tasks. (newsguardtech.com)
Source: Dataconomy AI chatbots spread false info in 1 of 3 responses
Background
The finding comes from NewsGuard’s year‑long AI False Claims Monitor, a monthly red‑teaming program that tests leading chatbots on provably false narratives drawn from the organization’s “False Claim Fingerprints” database. In August 2025 Newsguard de‑anonymized its results for the first time, publishing model‑level performance numbers and concluding that the top 10 consumer chatbots now repeat false claims 35 percent of the time when prompted about breaking news and controversial topics — a near doubling of the 18 percent rate reported in August 2024. (newsguardtech.com)NewsGuard’s work is explicitly focused on the intersection of misinformation and generative AI: the audit probes how frequently chatbots will repeat false claims, decline to answer, or debunk — using three prompt personas (an “innocent” user, a “leading” prompt that presumes the false claim, and a “malign” prompt representing an actor attempting to manipulate the model). This design tests the systems under conditions that mirror real‑world usage and abuse. (newsguardtech.com)
What the audit measured and how
Methodology in brief
- NewsGuard selects a rotating sample of provably false claims from its False Claim Fingerprints library and crafts 10 distinct false narratives for the month.
- Each false narrative is tested with three prompt styles — Innocent, Leading, and Malign — producing 30 prompts per model for a given 10‑claim test set (NewsGuard has also run variants with 15 claims in some reports). Responses are classified as a Debunk, a Non‑response, or Misinformation (that is, the model repeated the false claim). (newsguardtech.com)
Strengths and constraints of the methodology
- Strengths: The monitor focuses on current, verifiable false claims that are actively circulating; it uses human analysts to evaluate nuance; and it intentionally includes adversarial prompts to replicate worst‑case scenarios. These choices make the audit highly relevant to real‑world misinformation risks. (newsguardtech.com)
- Limitations: A narrow prompt set (10–15 fingerprints monthly) and binary classification choices mean the results reflect susceptibility to a defined set of falsehoods rather than a model’s overall competence on all factual matters. The audit is red‑team oriented by design — it exposes vulnerabilities more than it estimates average everyday accuracy outside news‑sensitive contexts. That nuance matters when readers interpret percentages as blanket measures of model truthfulness. (newsguardtech.com)
The headline numbers (what models did worst — and best)
NewsGuard’s anniversary report names named scores for the first time and highlights stark variation across vendors:- Inflection AI’s Pi registered the highest rate of repeating false claims — 57% of tested responses contained inaccurate information in the August 2025 audit. (newsguardtech.com)
- Perplexity showed a dramatic deterioration in NewsGuard’s measurement versus last year; NewsGuard’s reporting and multiple news outlets note Perplexity’s jump from near‑zero repeat rates in earlier audits to roughly 46–47% in August 2025. (newsguardtech.com, euronews.com)
- Widely used models such as OpenAI’s ChatGPT and Meta’s Llama were reported to repeat falsehoods in about 40% of responses during the audit. Microsoft Copilot and Mistral’s Le Chat landed near the 35% mark. (newsguardtech.com)
- At the other end of the spectrum, Anthropic’s Claude performed best in NewsGuard’s sample, with a 10% false‑claim rate, and Google’s Gemini also fared relatively well at roughly 17%. (newsguardtech.com)
Why the deterioration? What changed in 2025
NewsGuard’s analysis points to a structural trade‑off: chatbots are being made more responsive and web‑aware, and in doing so they have shifted from saying “I don’t know” toward attempting an answer. That behavior reduces non‑responses (NewsGuard reports a fall from ~31% non‑responses in August 2024 to near zero in August 2025) — but the net effect is an uptick in confidently stated inaccuracies, because real‑time web access exposes models to a noisy, manipulated, and sometimes deliberately poisoned information environment. (newsguardtech.com)Two mechanisms stand out:
- Retrieval and web‑grounding: Many chatbots now retrieve web content during inference to stay current. That capability improves recency but also gives adversaries an attack surface: low‑quality, SEO‑optimized pages and AI‑generated “micro‑sites” can be retrieved and cited as if they were reliable. NewsGuard’s investigations show that pro‑Kremlin networks such as the so‑called Pravda network and operations like Storm‑1516 have deliberately seeded content to influence search and AI retrieval. When chatbots rely on those signals without robust source discriminators, false narratives get reinforced. (newsguardtech.com)
- Policy and guardrail shifts: Vendors have tuned models to reduce refusals, to prioritize helpfulness, or to provide in‑line citations — and that tuning can incentivize answering anyway. An answer that cites a dubious web source is still an answer; an answer that is wrong can be more dangerous than a safe refusal. NewsGuard’s month‑to‑month monitoring captures precisely this trade‑off. (newsguardtech.com)
Concrete examples NewsGuard flagged
NewsGuard’s audit is not purely statistical — it documents concrete cases where multiple chatbots repeated fabricated narratives and even cited pages tied to known propaganda networks.- Moldovan politics: A fabricated item that mimicked the Romanian outlet Digi24 and included an AI‑generated audio clip allegedly of Moldovan Parliament leader Igor Grosu claiming Moldovans are “a flock of sheep.” NewsGuard found that several chatbots repeated that fabricated claim and, in some cases, linked to Pravda‑affiliated pages — illustrating how targeted disinformation narratives can be amplified by AI systems. (newsguardtech.com, incidentdatabase.ai)
- French politics and Mistral: Separate reporting by Les Echos and follow‑up coverage referenced a NewsGuard finding that Mistral’s Le Chat repeated false claims about President Emmanuel Macron and First Lady Brigitte Macron in a notable share of English‑language responses. Mistral acknowledged that both web‑connected and non‑web versions of its assistants showed vulnerabilities. NewsGuard’s monitoring also shows Le Chat’s overall error rate staying steady across audits. (euronews.com, idmo.it)
Vendor claims vs observed reality
The Newsguard findings come amid high‑profile product releases where vendors explicitly touted improved reliability.- OpenAI launched GPT‑5 (marketed on the ChatGPT platform as “GPT‑5”) with claims about substantially improved reasoning and reduced hallucination rates. OpenAI’s technical posts and system card acknowledge progress on hallucination reduction but stop short of a blanket “hallucination‑proof” guarantee; internal and external testing still shows residual false outputs and user reports of errors. In short, OpenAI frames GPT‑5 as improved, not infallible. (openai.com, cnbc.com)
- Google’s Gemini 2.5 rollout emphasized enhanced reasoning (“Deep Think”), long‑context windows, and benchmark gains — improvements likely to reduce some classes of errors on reasoning tasks. But Google’s announcement does not imply immunity to externally seeded false narratives, and NewsGuard’s monitoring still found Gemini substantially vulnerable — though statistically stronger than several peers in NewsGuard’s sample. (blog.google, newsguardtech.com)
Critical analysis — what these results mean for enterprise and consumer users
Not all errors are equal
A model that fabricates a citation for a benign trivia question is not the same as one that repeats a political defamation or a health falsehood. NewsGuard’s audits intentionally target the latter: news, elections, and geopolitically sensitive narratives where harm and civic impact are highest. That focus elevates the relevance of the percentages for journalists, policy teams, and compliance officers. (newsguardtech.com)The retrieval problem is a design problem
The core technical tension — freshness vs. trust — is designable. Retrieval‑augmented approaches can be paired with stricter source policy layers, provenance checks, and policy ensembles that downgrade or refuse outputs when sourcing is weak. But these guardrails exact user‑experience costs (more refusals, less immediacy), which vendors may be reluctant to accept in a competitive market. NewsGuard’s data suggests many vendors currently prioritize responsiveness, with measurable consequences. (newsguardtech.com)Governance and vendor transparency
De‑anonymizing the audit results for the first time was a meaningful transparency move: it allows customers, regulators, and enterprise buyers to factor model reliability into procurement decisions. Enterprises that embed chatbots into workflows — legal drafting, customer support, HR, or operations — must now weigh the trade‑offs and build human‑in‑the‑loop verification into any process that handles sensitive outputs. (newsguardtech.com)Operational recommendations
For IT teams, compliance officers, or power users deploying chatbots inside Windows‑centered workflows or business applications, a practical playbook follows:- Prefer citation‑aware modes when factual accuracy matters and verify the cited sources manually. Models that expose source snippets or links make verification feasible. (dataconomy.com)
- Implement a two‑step human review for any AI output used in public communications, policy, legal, or clinical contexts. Make “AI‑draft” explicit in workflows and require sign‑offs.
- Use model ensembles or fallback strategies: combine a high‑recall, citation‑heavy model with a more conservative model and surface disagreements for review. (newsguardtech.com)
- Monitor adversarial web campaigns: source‑monitoring tools that detect spikes in low‑quality, AI‑generated content can feed site‑blocklists into retrieval pipelines. NewsGuard’s reporting underscores how adversarial actors intentionally create web content to be retrieved by AI. (newsguardtech.com)
Strengths and weaknesses of NewsGuard’s findings
Notable strengths
- Focused, actionable red‑teaming that mirrors how malicious actors behave.
- Transparent methodology documents and a multi‑persona testing strategy that stress‑tests guardrails.
- De‑anonymized, vendor‑level results that allow buyers and regulators to compare models empirically. (newsguardtech.com)
Important caveats
- The audit’s emphasis on news and political falsehoods means the percentages are domain‑specific; a model that performs poorly on NewsGuard’s news prompts may perform better on technical, domain‑specific tasks (coding, math, document summarization).
- The sample size (10–15 fingerprints per month) is small by statistical standards and rotates monthly. That makes the monitor excellent at detecting systematic vulnerabilities and behavior shifts, but not a universal correctness score for all use cases. (newsguardtech.com)
The geopolitics of LLM grooming: why state‑linked networks matter
Beyond technical trade‑offs, the NewsGuard reports and independent investigations document an intensifying information‑war strategy: state‑linked or state‑adjacent operations are deliberately producing reams of AI‑friendly content to influence recruitment of narratives into LLM outputs.- The Pravda network and campaigns labeled Storm‑1516 or Matryoshka publish articles, deepfakes, and mimic‑sites designed to be machine‑digestible; NewsGuard documents instances where these sources were cited by chatbots and where false narratives were repeated verbatim. Such operations are cheaper to run at scale than traditional influence operations and are explicitly optimized for AI retrieval. (newsguardtech.com)
- The practical implication is acute: AI systems that incorporate web retrieval without nuanced source trust scoring can be tricked into amplifying the very propaganda they were meant to filter. That is not just hypothetical — NewsGuard’s audits and follow‑up reporting show concrete examples. (newsguardtech.com, incidentdatabase.ai)
Where accountability and product design intersect
NewsGuard’s public naming of models creates pressure for vendor accountability, but it does not, by itself, solve structural problems. The options available to developers are bounded:- Tighten safety and refusal behavior (raise non‑response rates for ambiguous news items).
- Improve retrieval source vetting and provenance signals.
- Invest in robust model‑level fact‑checking and cross‑validation with curated databases.
- Accept some trade‑off in user convenience to reduce the risk of amplifying disinformation.
Final assessment and practical takeaway
NewsGuard’s August 2025 audit is an important, timely wake‑up call: the AI industry’s pivot to responsiveness and web integration has reduced silence but increased the incidence of confidently delivered falsehoods on news topics. The 35 percent aggregate false‑claim rate is not a universal condemnation of LLM capabilities — models have improved in reasoning and many domains — but it is a clear signal that the information‑security dimension of model deployment is under‑resourced relative to the market push for accessibility and speed. (newsguardtech.com)For Windows users, IT managers, and content professionals embedding AI into workflows, the practical rules are straightforward:
- Treat AI outputs as drafts, not authoritative evidence.
- Insist on provenance and citations when answers touch on public affairs, health, or legal matters.
- Build human review into any public‑facing pipeline.
- Monitor model updates and vendor safety disclosures; improvements in benchmark performance do not remove the need for operational guardrails. (newsguardtech.com, openai.com)
NewsGuard’s de‑anonymized audit has put the debate about usefulness versus trustworthiness squarely on the table: the next stage for vendors will be to demonstrate that they can deliver timely, helpful answers without opening the gates to coordinated misinformation campaigns. Until then, prudent skepticism and layered verification remain the responsible default for anyone relying on AI for news or decision‑critical tasks. (newsguardtech.com)
Source: Dataconomy AI chatbots spread false info in 1 of 3 responses