Category Archives: AI

This is what happened when I asked journalism students to keep an ‘AI diary’

Last month I wrote about my decision to use an AI diary as part of assessment for a module I teach on the journalism degrees at Birmingham City University. The results are in — and they are revealing.

AI diary screenshots, including AI diary template which says:
Use this document to paste and annotate all your interactions with genAI tools. 

Interactions should include your initial prompt and response, as well as follow up prompts (“iterations”) and the responses to those. Include explanatory and reflective notes in the right hand column. Reflective notes might include observations about potential issues such as bias, accuracy, hallucinations, etc. You can also explain what you did outside of the genAI tool, in terms of other work. 

At least some of the notes should include links to literature (e.g. articles, videos, research) that you have used in creating the prompt or on reflecting on it. You do not need to use Harvard referencing - but the link must go directly to the material. See the examples on Moodle for guidance.

To add extra rows place your cursor in the last box and press the Tab key on your keyboard, or right-click in any row and select ‘add new row’.
Excerpts from AI diaries

What if we just asked students to keep a record of all their interactions with AI? That was the thinking behind the AI diary, a form of assessment that I introduced this year for two key reasons: to increase transparency about the use of AI, and to increase critical thinking.

Continue reading

How to reduce the environmental impact of using AI

Generative AI: reducing environmental impact
Disable AI or switch tool
Compare AI vs non-AI
Compare models
Prompt planning
Prompt design and templating
Measuring and reviewing
Run locally

One of the biggest concerns over the use of generative AI tools like ChatGPT is their environmental impact. But what is that impact — and what strategies are there for reducing it? Here is what we know so far — and some suggestions for good practice.

What exactly is the environmental impact of using generative AI? It’s not an easy question to answer, as the MIT Technology Review’s James O’Donnell and Casey Crownhart found when they set out to find some answers.

“The common understanding of AI’s energy consumption,” they write, “is full of holes.”

Continue reading

9 takeaways from the Data Journalism UK conference

Attendees in a lecture theatre with 'data and investigative journalism conference 2025 BBC Shared Data Unit' on the screen.

Last month the BBC’s Shared Data Unit held its annual Data and Investigative Journalism UK conference at the home of my MA in Data Journalism, Birmingham City University. Here are some of the highlights…

Continue reading

Teaching journalism students generative AI: why I switched to an “AI diary” this semester

The Thinker status
Image by Fredrik Rubensson CC BY-SA 2.0

As universities adapt to a post-ChatGPT era, many journalism assessments have tried to address the widespread use of AI by asking students to declare and reflect on their use of the technology in some form of critical reflection, evaluation or report accompanying their work. But having been there and done that, I didn’t think it worked.

So this year — my third time round teaching generative AI to journalism students — I made a big change: instead of asking students to reflect on their use of AI in a critical evaluation alongside a portfolio of journalism work, I ditched the evaluation entirely.

Continue reading

Why I’m no longer saying AI is “biased”

TLDR; Saying “AI has biases” or “biased training data” is preferable to “AI is biased” because it reduces the risk of anthropomorphism and focuses on potential solutions, not problems.

Searches for "AI bias" peaked in 2025. In March 2025 twice as many searches were made for "AI bias" compared to 12 months before.
Click image to explore an interactive version

For the last two years I have been standing in front of classes and conferences saying the words “AI is biased” — but a couple months ago, I stopped.

As journalists, we are trained to be careful with language — and “AI is biased” is a sloppy piece of writing. It is a thoughtless cliche, often used without really thinking what it means, or how it might mislead.

Because yes, AI is “biased” — but it’s not biased in the way most people might understand that word.

Continue reading

How do I get data if my country doesn’t publish any?

Spotlight photo by Paul Green on Unsplash

In many countries public data is limited, and access to data is either restricted, or information provided by the authorities is not credible. So how do you obtain data for a story? Here are some techniques used by reporters around the world.

Continue reading

7 prompt design techniques for generative AI every journalist should know

Tools like ChatGPT might seem to speak your language, but they actually speak a language of probability and educated guesswork. You can make yourself better understood — and get more professional results — with a few simple prompting techniques. Here are the key ones to add to your toolkit. (also in Portuguese)

Prompt design techniques for genAI
Role prompting
One-shot prompting
Recursive prompting
Retrieval augmented generation
Chain of thought
Meta prompting
Negative prompting
Continue reading

What are we talking about when we say the “ethics” of AI and journalism?

A path forking into three
There are three broad paths in ethics. Image by pfly CC BY-SA 2.0

Many people — including me — are quite uncomfortable with generative AI. Most of this discomfort can be traced to the various ethical challenges that AI raises. But an understanding of the different schools of ethics can help us both to better address those challenges and what to do about them.

Three different ethical approaches

The first thing to say about the ethics of AI is that there is no single ‘ethics’. When we engage with ethical issues there are typically at least three different systems that might be in play:

Continue reading

Identifying bias in your writing — with generative AI

Applications of genAI in the journalism process 
Pyramid with the third 'Production' tier highlighted: 

Identify jargon and bias; improve spelling, grammar, structure and brevity

In the latest in a series of posts on using generative AI, I look at how tools such as ChatGPT and Claude.ai can help help identify potential bias and check story drafts against relevant guidelines.

We are all biased — it’s human nature. It’s the reason stories are edited; it’s the reason that guidelines require journalists to stick to the facts, to be objective, and to seek a right of reply. But as the Columbia Journalism Review noted two decades ago: “Ask ten journalists what objectivity means and you’ll get ten different answers.”

Generative AI is notoriously biased itself — but it has also been trained on more material on bias than any human likely has. So, unlike a biased human, when you explicitly ask it to identify bias in your own reporting, it can perform surprisingly well.

It can also be very effective in helping us consider how relevant guidelines might be applied to our reporting — a checkpoint in our reporting that should be just as baked-in as the right of reply.

In this post I’ll go through some template prompts and tips on each. First, a recap of the rules of thumb I introduced in the previous post.

Continue reading