Should the media tell you when they use AI to report the news? What consumers should know

Should the media tell you when they use AI to report the news? What consumers should know

Artificial intelligence (AI) is transforming the media landscape, both for news organisations and consumers. Applications such as ChatGPT, Bard, and Bing AI are creating new possibilities to assist in writing and researching the news, but these also raise ethical concerns.

One of the most pressing questions for news organisations is whether consumers should be told when they are reading a story created, or aided by, use of AI. Some, such as the technology magazine Wired and the BBC are already doing this, but other media outlets are not.

There are several arguments for and against disclosing this kind of information.

First, it would help to ensure transparency and accountability. Consumers should know how the news they are consuming is being produced, and they should be able to make informed choices about whether or not to trust it.

Second, disclosure could help mitigate the risks of bias. AI systems are trained using data, and that data can reflect the biases of the people who created it. As a result, AI-generated content can sometimes be biased. By requiring disclosure, consumers would be able to be aware of this potential bias and take it into account when evaluating the information.

Third, disclosure could help to protect consumers from misinformation . AI systems can be used to generate fake news, making it difficult for consumers to distinguish between real and fake news. By requiring disclosure, consumers would be able to be more sceptical of AI-generated content and be more likely to verify it before sharing it.

Against disclosure

One concern is that it could stifle innovation. If news organisations are required to disclose every time they use AI, they may be less likely to experiment with the technology.

See also  Thought-provoking new exhibition suggests the public should help shape the future of AI

Another is that disclosure could be confusing for consumers. Not everyone understands how AI works. Some people may be suspicious of AI-generated content. Requiring disclosure could make it more difficult for consumers to get the information they need.

How things could play out

Here are a couple of examples to illustrate these concerns:

Imagine a news organisation is using AI to perform real-time fact-checking and verification of statements made by public figures during live events, such as political debates or press conferences. An AI system could rapidly identify inaccuracies and provide viewers with accurate information in real-time.

However, if the news organisation were required to disclose the use of AI each time, it might lead to a reluctance to deploy such a tool. The fear of public perception and potential backlash could deter news outlets from leveraging AI to enhance the accuracy of their reporting, ultimately depriving the audience of a valuable service.

Another scenario involves AI-driven personalised news curation. Many news platforms use AI algorithms to tailor news content to individual readers’ preferences, ensuring they receive information that aligns with their interests.

If news organisations were compelled to disclose the use of AI in this context, readers might become wary of perceived manipulation. This apprehension could deter news outlets from investing in AI-driven personalisation, limiting their ability to engage and retain audiences in an increasingly competitive media landscape.


Read more:
Deepfakes in warfare: new concerns emerge from their use around the Russian invasion of Ukraine

To mitigate these risks, publications such as the New York Times are offering “enhanced bylines” that include more details about the journalists behind the stories and details about how the story was produced.

See also  Digital privacy and Covid-19: between a paradox and a hard place

Ultimately, the decision of whether or not to require disclosure is a complex one.

However, it is essential to have a public conversation about this issue so that we can develop policies that protect consumers and promote responsible journalism, and retain and improve trust in journalism, which is falling in some countries.

In addition to disclosure, there are other things that news organisations can do to ensure that AI is used ethically and responsibly. They should develop clear guidelines for the use of AI. These guidelines should address issues such as bias, transparency and accountability. They should invest in training and education for their staff. Journalists need to understand how AI works and how to use it responsibly.

Finally, news organisations should work with highly informed groups such as Harvard’s Neiman Lab, those working on policy, technology companies and academics, to develop ethical standards for using AI and tackle emerging issues critical to the future of public-interest news.

The use of AI tools in news is a significant development . It is vital to have a thoughtful and informed conversation about this technology’s potential benefits and risks. By working together, we can ensure that AI is used in a way that serves the public interest and upholds the values of responsible journalism.

François Nel is an editor of the World Press Trends, the flagship report of WAN-IFRA, the World Association of News Publishers, and he receives funding from the Google News Initiative for the News Futures 2035 project, which is a multi-stakeholder foresight study into the sustainable supply of trustworthy public interest news in the UK.