Will artificial intelligence be the death of journalism?

In mid-May 2023, The Irish Times published an article that accused women who use fake tan of mocking those with naturally dark skin. The op-ed was initially said to be written by Adriana Acosta-Cortez, a 29-year-old Ecuadorian health worker living in north Dublin.

But no such person existed. “The article and the accompanying byline photo may have been produced, at least in part, using generative AI technology,” read an editorial in The Irish Times four days after the piece first published.

Two months later, HoldtheFrontPage— a news website for journalists with a focus on regional media across the UK— published an investigative piece documenting how artificial intelligence (AI) was used to launch a publication purporting to be called The Bournemouth Observer, which turned out to be a fake newspaper. “It was obvious that the content was written by AI because the writing was so bad,” editor of HoldtheFrontPage, Paul Linford, told Index. “But since then, AI has got much better at writing stories, and I suspect it will eventually become harder to spot when writing is being done by AI or real journalists,” said Linford.

Index on Censorship was also caught out by a journalist calling themselves Margaux Blanchard whose article was published in the Spring edition of the magazine. Ironically it was about journalism in Guatemala and written by AI.  Others – Wired and Business Insider – also fell victim to “Margaux”.

James Barrat claimed AI “will eventually bring about the death of writing as we know it.” The American documentary maker and author has been researching and writing about AI for more than a decade. His previous books include Our Final Invention: Artificial Intelligence and the End of the Human Era (2013), which ChatGPT recently ingested. “There are presently ongoing lawsuits about this because OpenAI took my book [without my permission] and didn’t pay me,” Barrat explained. “Right now, if you tell ChatGPT ‘write in the style of James Barrat’ it doesn’t produce an exact replica, but it’s adequate, and machine writing is getting better all the time.”

AI “will eliminate 30% of all jobs”

In early September, Barrat published The Intelligence Explosion: When AI Beats Humans at Everything (2025). The book makes two bold predictions. First, AI has the potential in the not-too-distant future to potentially match, and perhaps even surpass, our species’ intelligence. Second, by 2030 AI will eliminate 30 percent of all jobs done by humans, including writers. Freelance journalists will benefit in the short term, Barrat claimed. “Soon a basic features writer, using AI, will be able to produce twice as much content and get paid twice as much,” he said. “But in long run the news organisations will get rid of [most] writers because people won’t care if content is written by AI or not.”

Tobias Rose-Stockwell did not share that view. “There will always be a market for verified accurate information, which requires humans,” the American writer, designer, technologist and media researcher said. “So truthful journalism isn’t going away, but it’s going to be disrupted by AI, which can now generate content in real time. This will lead to more viral falsehoods, confusion and chaos in our information ecosystem.”

Rose-Stockwell elaborated on this topic in Outrage Machine (2023). The book documents how the rise of social media in the mid-2000s was made possible by algorithms, which are mathematical instructions that process data to produce specific outcomes. In the early days of social media users viewed their feeds in chronological order. Eventually, though, Facebook, Instagram, X, TikTok, and other social media platforms realised it was more profitable to organise that information via algorithmic feeds, powered by artificial intelligence and, in particular, machine learning (ML) where AI is used to identify behaviours and patterns that may be missed by human observers. ML tools analyse users’ behaviour, preferences, and interactions, keeping them emotionally engaged for longer. “Feed algorithms are much better at re-coordinating content than any human ever could,” said Rose-Stockwell. “They can even create bespoke little newspapers or television shows for us.”

“AI is already in the process of rapidly transforming journalism,” said Dr Tomasz Hollanek, a technology ethics specialist at the University of Cambridge with expertise in intercultural AI ethics and ethical human-AI interaction design. “As AI systems become more adept at producing content that appears authentic, detecting fabricated material will get harder.”

Hollanek spoke about editors giving journalists clear guidelines about when and where AI can be used. The Associated Press, for instance, currently allows staff to test AI tools but bans publishing AI-generated text directly.

“What’s important about these guidelines is that while they recognise AI as a new tool, they also stress that journalism already has mechanisms for accountability,” said Hollanek. He also criticised the sensationalist tone journalists typically take when writing about AI, pointing to unnecessary hype, which leads to distorted public understanding and skewed policy debates.

“Journalists strengthening their own critical AI literacy will make the public more informed about AI and more capable of shaping its trajectory.”

AI’s role in news “needs to be trackable”

Petra Molnar, a Canadian lawyer and anthropologist who specialises in migration and human rights, claimed “the general public needs to understand that AI is not some abstract tool out there, but it’s already structuring our everyday lives.”

Molnar said there is an urgent need for public awareness campaigns that make AI’s role in news and politics visible and trackable. She described companies such as Meta, X, Amazon, and OpenAI as “global gatekeepers [of information] with the power to amplify some voices while silencing others, often reinforcing existing inequalities.”

“Most people experience AI through tools like news feeds, predictive texts, or search engines, yet many do not realise how profoundly AI shapes what they see and think,” said Molnar, who is the associate director of the Refugee Law Lab at York University, Toronto— which undertakes research and advocacy about legal analytics, artificial intelligence, and new border control technologies that impact refugees.“AI is often presented as a neutral tool, but the reality is that it encodes power.”

Last year, Molnar published The Walls Have Eyes: Surviving Migration in the Age of Artificial Intelligence (2024). The book draws attention to a recent proliferation across the globe of digital technologies that are used to track and survey refugees, political dissidents, and frontline activists crossing borders in times of conflict. Molnar claimed that “AI threatens to accelerate the collapse of journalism by privileging speed and engagement over accuracy and depth.” She cited examples of journalists using OpenAI-generated text tools to churn our surface-level articles that echo sensational framings around migration, without investigative depth.

“Automated systems may generate content that looks like journalism, but it’s stripped of accountability and critical inquiry that’s required to tell complex stories,” said Molnar. “Journalism’s future depends on human reporters who can investigate power and rigorously fact check, something AI simply cannot replicate.”

Human oversight

Sam Taylor, campaigns & communications officer at the UK’s National Union of Journalists (NUJ), shared that view. “Editors and writers should exercise caution before using AI in their work,” he said. “Generative AI often draws on databases from the internet that contains stereotypes, biases, and misinformation.”

“To maintain and strengthen public trust in journalism, AI must only be used as an assistive tool with human oversight,” said NUJ general secretary, Laura Davison.

Everyone Index spoke to agreed that AI, for all its flaws, offers journalists enormous benefits, including automating mundane routine tasks, like transcription and data sorting. AI can also make data journalism, exploring large data sets to uncover stories, much more accessible as AI can crunch the data and identify interesting nuggets far faster than a person can. This will leave journalists with more time and energy for critical thinking, and ultimately, to tell more complex and nuanced stories.

But there was also an overwhelming consensus that AI cannot fact-check accurately or be trusted as a credible verifier of information. Not least because it suffers from hallucinations. “This means due to the complexity of what is going on inside it hallucinates,” James Barrat explained. “When this happens, AI gets confused and tells lies.”

“The jury is still out on whether or not this hallucination problem can be solved or not,” said Tobias Rose-Stockwell. “Journalism must remain grounded in ethical responsibility and context,” said Petra Molnar. “What we need is human judgement, applied critically and ethically, supported by but not replaced by technology.”

Is AI a threat?

Anyone who believes in journalism’s primary mission, to challenge power by investigating the truth, is undoubtedly likely to agree. But is this wishful thinking from a bygone era? James Barrat believes so. He points out that, eventually, we may not have the option to choose. “A scenario that could happen in near future is that AI could become hostile to us,” he said. “AI could take control of our water and our electrical systems. Just recently, a large language model (LLM) agreed that its creator should be killed.”

Barrat mentions an interview he did with the British science fiction writer and futurist Sir Arthur C Clarke, before his death, aged 90, in 2008. Clarke co-wrote the screenplay for 2001: A Space Odyssey (1968). Directed by Stanley Kubrick, the Oscar winning film tells the story of an AI-powered computer— aboard the Discovery One spacecraft bound for the planet Jupiter— called HAL. Eventually, HAL experiences a program conflict, malfunctions, and, to defend itself, turns on its human crew members.

Arthur C Clarke told Barrat, “Humans steer the future because we are the most intelligent species on the planet, but when we share the planet with something smarter than us, they will steer the future.”

“AI’s intelligence grows exponentially,” Barrat concludes. “As it gets smarter, we will stop understanding it. We really are inviting disaster.”

The autumn issue of Index magazine, titled Truth, trust and tricksters and published on 26 September, looks at the threats that artificial intelligence poses to freedom of expression

From Survivor to Defendant: How the law is being weaponised to silence victims of sexual violence

Since 2017, the #MeToo movement has encouraged millions of women to share their experiences of sexual and gender-based violence (SGBV). But speaking out hasn’t come without risks. Some survivors (and the journalists covering their stories) have faced legal threats from the very people accused of perpetrating SGBV. As a result, British courts have been weaponised to continue abuse, with few protections in place.

SLAPPs (strategic lawsuits against public participation) are lawsuits aimed at stifling public interest speech. They are designed not to win in court, but to intimidate, drain resources and silence voices. Their aim isn’t justice, but suppression.

Join us for the launch, sponsored by Baroness Helena Kennedy KC, of From Survivor to Defendant, a new report from Index on Censorship exploring how SLAPPs are being used against survivors of SGBV across the UK and Ireland. 

With words from Baroness Helena Kennedy KC (House of Lords), Jemimah Steinfeld (Index on Censorship), Olga Jurasz (Centre for Protecting Women Online, The Open University), SLAPP targets Nina Cresswell and Verity Nevitt, and Mark Stephens CBE (Howard Kennedy).

RSVP to confirm your attendance here.

The report was prepared with the generous support of the Joseph Rowntree Charitable Trust and the Open University’s Centre for Protecting Women Online.

Index is grateful for the contributions of the survivors, and especially the Gemini Project who shared their research with us.

We are grateful to our event co-sponsors, the International Bar Association’s Human Rights Institute.

The EU is losing patience with Hungary

The European Convention on Human Rights was set up in the aftermath of World War Two to protect the rights of people in the Council of Europe’s 47 member states. Enshrined within it are fundamental obligations around free speech, including the right to free expression and the right to protest. It was intended to act as a blueprint for democracy and a rules-based order – but certain member states are tearing up this rulebook, none more so than Viktor Orbán’s Hungary, and seemingly getting away with it.

In recent years, Orbán has intensified his crackdown on democratic principles, including eroding academic freedoms and increasing hostility towards the media. In April, Orbán even decided to withdraw Hungary from the International Criminal Court – a pan-global organisation set up to uphold the rule of law and hold those charged with the gravest war crimes accountable.

The LGBTQ+ community has been particularly targeted in Hungary, with laws passed abolishing the legal recognition of transgender people in 2020 and banning the depiction of homosexuality to under-18s in 2021. A recent escalation is a law banning Pride marches, introduced in March, under the guise that such gatherings are harmful to children. At the time, Orbán said: “We won’t let woke ideology endanger our kids.”

Hungary’s parliament has since passed a series of other amendments tightening the government’s grip on those seeking to attend Pride, which will allow authorities to use facial recognition technology to identify people at events, and potentially fine them up to 200,000 Hungarian forint (HUF), the equivalent of $560. Protests have erupted across Hungary since the law was passed, and thousands are expected to turn out in defiance at Budapest Pride on 28 June.

Meanwhile, other draft legislation is making its way through parliament that is reminiscent of Russia’s “foreign agent” law – the Transparency of Public Life bill, if passed, would allow the government to penalise and ban dissenting voices and critics deemed detrimental to Hungary’s national interests, including the press and NGOs.

Hungary’s recent actions not only contravene the ECHR, but also the European Union’s (EU) policies around democracy and human rights, as laid out in its treaties and the EU Charter of Fundamental Rights, of which its 27 member states are meant to abide by.

But the political tide appears to be turning. Hungary is clearly moving further and further away from the fundamental values of the EU, and many member countries are growing frustrated with the Central European state, and with the European Commission for not taking strong enough action.

This week, 17 EU countries, including France, Germany, Ireland and Spain, signed a declaration expressing their concerns and dismay over Hungary’s anti-Pride law. They have called on Orbán to revise it, and have asked the European Commission to take legal action against Hungary if it does not do so.

Michael McGrath, the EU commissioner responsible for democracy, said this week the “willingness is there” to take action against Hungary, and that a “comprehensive analysis of the relevant legislation is underway now”.

But so far, retribution for Hungary’s actions has been negligible. The European Council has discussed Hungary’s rule of law violations seven times in the European Parliament since 2018, but has never taken the next step in the process, which would allow member states to vote on sanctions against Hungary.

There have been some financial penalties, but relations between the EU and Hungary are likely complicated by the need for cooperation against Russia’s full-scale invasion of Ukraine. In December 2022, the European Commission froze €30 billion ($34 billion) in funds to Hungary, after the country’s failure to address concerns around democracy and the rule of law; a year later, a third of these funds were unfrozen, with speculation that Orbán was threatening to impede the EU’s actions in supporting Ukraine.

Patience with Hungary amongst EU nations appears to be wearing thin. The question is: how long will Orbán’s impunity be allowed to continue, and what example does this set for other EU countries wishing to replicate his methods?

AIFAS methodology

Published methodology The new Am I Facing a SLAPP tool has been developed by Index on Censorship with funding from the Justice For Journalists Foundation and the Joseph Rowntree Charitable Trust. It is based on an earlier version of the tool launched in 2020, which...
SUPPORT INDEX'S WORK