Women represent 49.5% of the world’s population, but they do not have a corresponding public, political and social influence. In recent years, more and more women have raised their voices, making society aware of their challenges — data journalists included. To commemorate International Women’s Day, Carla Pedret presents a list of data journalism projects that detail the sacrifices, injustices and prejudices that women still have to face in the 21st century.
In a guest post for OJB, Barbara Maseda looks at how the media has used text-as-data to cover State of the Union addresses over the last decade.
State of the Union (SOTU) addresses are amply covered by the media —from traditional news reports and full transcripts, to summaries and highlights. But like other events involving speeches, SOTU addresses are also analyzable using natural language processing (NLP) techniques to identify and extract newsworthy patterns.
Every year, a new speech is added to this small collection of texts, which some newsrooms process to add a fresh angle to the avalanche of coverage.
If you’ve ever dug a story out of a pile of text, you can contribute to Barbara’s research by answering a few questions here.
In a guest post for OJB, Carla Pedret looks at a new data journalism project to catalogue what happened during the Spanish Civil War.
125,000 people died, disappeared or were repressed in the Spanish Civil War (1936-1939) and during the Franco dictatorship, according to historians. Many of their families still do not know, 40 years later, what exactly happened to them.
When I agreed to write the second edition of the Online Journalism Handbook, I asked that the chapter on data journalism from the 2011 edition of the book be released under a Creative Commons licence. To Routledge’s credit, they agreed. Here, then, I’m making that book chapter available — you can download it from here or access it on Slideshare (embedded below).
It’s always difficult to get publishers to agree to things like this, so if you have any comments or feedback that I can use to make a similar case to publishers in future, please let me know in the comments.
- You can read about the changes between the first and second editions of the Online Journalism Handbook in this post.
The work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License.
The latest in my series of FAQ posts comes from the National University of Sciences & Technology (NUST) in Pakistan. As always, I’m publishing my answers to their questions here in case it’s of use to anyone else.
Q. What would you say to convince journalists — especially journalists working in developing countries where even the acquisition of public records is often a tedious task — about the importance of data journalism?
If you believe that journalism has a duty to be factual, accurate, and to engage an audience in subjects which have a clear public and civic importance, then data journalism is going to be very important to your work. Continue reading
The prices of my 3 data journalism ebooks — Data Journalism Heist, Finding Stories in Spreadsheets and Scraping for Journalists — have been cut to $5 on Leanpub in the lead up to Christmas. And if you want to get all 3, you can also get the data journalism books bundle on Leanpub for more than half price over the same period, at $13. Get them while it lasts!
Last year I was part of a team — with Yemisi Akinbobola and Ogechi Ekeanyawu — that won a CNN MultiChoice African Journalist of the Year award for an investigation into Nigerian football agents. The project, funded by Journalismfund.eu, and also available in an immersive longform version, combined data journalism and networked production with on-the-ground reporting. Here are some of the lessons we drew from the project… Continue reading
Last week I hosted the second annual Data Journalism UK conference — a convenient excuse to bring together speakers from the news media, industry experts, charities and startups working in the field. You can read write ups on Journalism.co.ukand the BBC Academy website (who kindly hosted the event at BBC Birmingham), but I thought I’d also put my own thoughts down here…
The Bureau and the BBC: 2 networked models for supporting data journalism
2017 saw the launch of two projects with a remit to generate and stimulate data journalism at a local level: the Bureau of Investigative Journalism’s Bureau Local project, and the BBC’s Shared Data Unit. Continue reading
Earlier this year I announced a new MA in Data Journalism. Now I am announcing a version of the course for those who wish to study a shorter, part time version of the course.
The PGCert in Data Journalism takes place over 8 months and includes 3 modules from the full MA:
- Data Journalism;
- Law, Regulation and Institutions (including security); and
- Specialist Journalism, Investigations and Coding
The modules develop both a broad understanding of a range of data journalism techniques before you choose to develop some of those in greater depth on a specialist project.
The course is designed for those working in industry who wish to gain accredited skills in data journalism, but who cannot take time out to study full time or may not want a full Masters degree (a PGCert is 60 credits towards the 180 credits needed for a full MA).
In a special guest post Anders Eriksen from the #bord4 editorial development and data journalism team at Norwegian news website Bergens Tidende talks about how they manage large data projects.
Do you really know how you ended up with those results after analyzing the data from Public Source?
Well, often we did not. This is what we knew:
- We had downloaded some data in Excel format.
- We did some magic cleaning of the data in Excel.
- We did some manual alterations of wrong or wrongly formatted data.
- We sorted, grouped, pivoted, and eureka! We had a story!
Then we got a new and updated batch of the same data. Or the editor wanted to check how we ended up with those numbers, that story.
…And so the problems start to appear.
How could we do the exact same analysis over and over again on different batches of data?
And how could we explain to curious readers and editors exactly how we ended up with those numbers or that graph?
We needed a way to structure our data analysis and make it traceable, reusable and documented. This post will show you how. We will not teach you how to code, but maybe inspire you to learn that in the process. Continue reading