Home » AI chatbots are wrong about news 45% of the time, study finds

AI chatbots are wrong about news 45% of the time, study finds

by Nia Walker
2 minutes read

AI Chatbots: Addressing the 45% Error Rate in News Delivery

In a recent international study spearheaded by the European Broadcasting Union (EBU) and spearheaded by the BBC, a startling revelation emerged regarding AI chatbots and their handling of news content. The study, encompassing input from 22 public service broadcasters across 18 countries and 14 languages, highlighted a significant concern: AI assistants are getting it wrong about the news nearly half of the time.

Professional journalists meticulously scrutinized over 3,000 AI responses from platforms like Chat GPT, Copilot, Gemini, and Perplexity. The findings were alarming, with 45% of the responses containing serious errors, 31% lacking proper or accurate source citations, and 20% riddled with major factual inaccuracies, including fabricated information or outdated details.

Notably, Google’s Gemini emerged as the poorest performer, with a staggering 76% of its responses marred by issues primarily stemming from a lack of proper source attribution. EBU Media Director Jean Philip De Tender emphasized the systemic nature of these flaws, underscoring their widespread impact on public trust. The study’s implications are profound, suggesting that a lack of credibility in AI-generated news content could erode democratic participation by fostering a climate of distrust.

In response to these findings, the EBU and the BBC have taken proactive steps by introducing the “News Integrity in AI Assistants Toolkit.” This toolkit aims to assist AI developers and users in enhancing the quality of responses while promoting media literacy among consumers. Additionally, both organizations have called upon the European Union and national regulatory bodies to enforce existing regulations on information integrity, digital services, and media diversity. They have also advocated for the implementation of independent oversight mechanisms to evaluate AI assistants continually.

Despite the gravity of these revelations, major players in the AI landscape such as Open AI, Microsoft, Google, and Perplexity AI have yet to provide feedback on the study’s outcomes. This underscores the importance of industry-wide collaboration and accountability in addressing the challenges posed by AI chatbots in news dissemination.

As technology continues to evolve rapidly, ensuring the accuracy and reliability of AI-generated content remains a pressing concern. By acknowledging the shortcomings unearthed in this study and collectively working towards enhancing the integrity of AI assistants, we can uphold the principles of transparency and credibility in the digital age. Let us seize this opportunity to foster a more informed and trustworthy news ecosystem for the benefit of all stakeholders involved.

You may also like