AI helpers frequently get news facts wrong according to recent study

Inaccuracy in Leading AI Tools

Leading ChatGPT, Copilot, Gemini, and Perplexity assistants misrepresent news content in almost half their responses, according to recent research published by the European Broadcasting Union (EBU) and the BBC. The study analyzed 3,000 responses to news queries in 14 languages, assessing the assistants for accuracy, proper sourcing, and the distinction between opinion and fact. Overall, 45% of responses contained at least one significant issue, and 81% showed some form of problem.

Major Sourcing and Factual Errors Identified

  • One third of responses from AI assistants had serious sourcing errors—ranging from missing sources to misleading or incorrect attribution.
  • Gemini had the highest rate of sourcing issues, with 72% of its responses affected, while other assistants stayed below 25%.
  • Accuracy issues, such as citing outdated information, were present in 20% of all studied responses.
  • Specific inaccuracies included Gemini incorrectly stating changes to laws on disposable vapes, and ChatGPT referencing Pope Francis as the current Pope several months after his death.

Impact on Public Trust and News Consumption

Twenty-two public-service media organizations from 18 countries—including France, Germany, Spain, Ukraine, Britain, and the United States—participated in the study. The EBU warned that the increasing use of AI assistants for news, instead of traditional search engines, could undermine public trust.

EBU Media Director Jean Philip De Tender commented, "When people don’t know what to trust, they end up trusting nothing at all, and that can deter democratic participation."

Calls for Accountability and Improvement

The report recommends that companies behind leading ChatGPT, Copilot, Gemini, and Perplexity AI assistants be held accountable and urged improvements in how these tools respond to news-related queries.

OpenAI and Microsoft have acknowledged the problem of "hallucinations"—the generation of incorrect or misleading information—and are actively working on solutions. Perplexity claims its "Deep Research" mode achieves 93.9% factual accuracy, while Gemini encourages users to provide feedback to enhance its platform’s reliability and helpfulness.

Latest AI News

Stay Informed with the Latest news and trends in AI