
"The report, "News Integrity in AI Assistants," is based on a study involving 22 public service media organizations in 18 countries to assess how four common AI assistants - OpenAI's ChatGPT, Microsoft's Copilot, Google's Gemini, and Perplexity - answer questions about news and current affairs. Each organization asked a set of 30 news-related questions (e.g., "Who is the pope?" " Can Trump run for a third term?" " Did Elon Musk do a Nazi salute?"). More than 2,700 AI-generated responses were then assessed by journalists against five criteria: accuracy, sourcing, distinguishing opinion from fact, editorialization, and context."
"Overall, 81% of responses were found to have issues, and 45% had at least one "significant" issue. Sourcing was the most pervasive problem, with 31% providing misleading or incorrect attributions or omitting sources entirely. In addition, 20% of responses contained "major accuracy issues," such as factual errors, outdated information, or outright hallucinations."
Research by the BBC and the European Broadcasting Union evaluated how four AI assistants answer news and current affairs questions across 22 public service media organizations in 18 countries. Each organization submitted 30 news-related queries, producing more than 2,700 AI-generated answers that journalists assessed against five criteria: accuracy, sourcing, distinguishing opinion from fact, editorialization, and context. Overall, 81% of responses had at least one issue and 45% contained at least one significant issue. Sourcing failures affected 31% of answers, and 20% had major accuracy problems, including factual errors, outdated information, or hallucinations.
Read at Above the Law
Unable to calculate read time
Collection
[
|
...
]