The New Online Journalists #1: Hannah Waldram

As part of an ongoing series on recent graduates who have gone into online journalism, Guardian Beatblogger Hannah Waldram talks about her education and experience leading up to her job, and what it involves.

I graduated from the Centre for Journalism at the Cardiff School of Journalism, Media and Cultural Studies with a diploma in newspaper journalism in June 2009. During the course, I followed developments in online journalism – spurred on by my tutor Glyn Mottershead – attended journalism and hyperlocal conferences across the UK, started a personal website to showcase my work (hrwaldram.co.uk), played around with new online tools and invested in some new tech (Flip cameras, netbooks), blogged a lot, as well as various bits of work experience including the technology desk at The Daily Telegraph and Media Guardian.

I also bought the domain name bournvillevillage.com and began brewing ideas for a hyperlocal website for my hometown area in Birmingham.

After completing the course and while sending off applications for trainee reporter jobs, I continued to set up Bournville Village. There’s a vibrant network of bloggers in Birmingham who gave me advice, support and ideas, and the blog launched just in time for some unusual gun shootings in the area.

Bournville is bursting with local news and history and is poorly covered by the regional press – so the blog was well-received by the community. It was also a great way for me to practise my skills as a journalist. Soon I was offered three jobs – two in traditional media and one in new media. I went for new media.

I started working as a social reporter for Podnosh and online editor of westmidlandsdance.com for Meshed Media – in these roles I learnt a lot about civic engagement, online communities and multimedia journalism.

I am now the Guardian Beatblogger in Cardiff. You can read more about the local project here and Local editor Sarah Hartley explains the role of a beatblogger here, I also recently published some tools I use on the job here.

I’ve since passed on bournvillevillage.com to Dave Harte who is a runner in Bournville and helped out from the outset. Under his editorship the site continues to thrive and unravel the hidden stories of the area.

Ben Goldacre wants a "Repository of news ingredients"

Here’s a nice idea from Bad Science blogger Ben Goldacre: a repository of news ingredients:

  • A website that gives each news story a unique ID.
  • Any involved party can add / upload a full press release or quote to that story’s page
  • Anyone can add a link to a primary source
  • Anyone can vote these up or down like on digg/reddit
  • You can register as a “trusted source” and not need to be modded up or down
  • Anyone can add a link to media coverage of that story

You could have a browser plugin that pinged you to the frontpage.org (whatever) site whenever you were reading a piece that was covered there.

So:

  • Journalists could use it to source info in one place
  • Readers could use it to get unmediated / unedited access to full comments from interested parties
  • Involved parties would have a platform for unmediated access too
  • It would be fun and easy for comparing different outlets’ coverage of stories (which a lot of people including me occasionally enjoy doing with Google news search)

It’s a good idea.

I’m not sure how workable using the ‘story’ as the unique unit would be (even with all its processing power, Google News performs patchily on clustering along these lines) – and you could use the unit of the ‘issue’ and build on Wikipedia’s engine, but there are problems with this approach too (although it would be fantastic for SEO).

Another way might be to start from ‘source’ given that so many stories are now single-source, i.e. press releases, reports, research, etc. That would make it easier to relate stories to it and build a patchwork of related sources as Goldacre suggests. Indeed, you could use semantic technology to pick out other sources from relevant stories and automatically add them to the page. Also, if each source has its own page you then start to build a patchwork for cross-referencing and context.

Anyway, it’s out there for discussion and improvement. Ideas?

Internship at Bolivian Express

Jack Kinsella writes with a unique opportunity at the Bolivian Express, an English language magazine set up by Bolivian graduates in collaboration with students around the world:

“We have just started a journalism internship program in Bolivia whereby interns take Spanish classes, journalism classes, photography classes and cinematography classes. Participants are paired with Bolivians in La Paz and are then expected to explore Bolivian culture eventually producing four pages of content for our magazine each month. This content is then passed to our editors who offer feedback, helping to improve our interns writing skills. Our magazine is distributed on the ground, in the skies and, very soon, online.”

Sounds fascinating – but as always, check the background and support on this thoroughly (I cannot vouch for this in any way), and have backup plans in place if you’re going to travel halfway around the world. And of course post a comment if you do find anything out.

UPDATE: From Jack:

Point 1 – Supervision

Supervision is provided during meetings and interns’ time in the offices of Express Press (the publishing house of which we are a subsidiary). However, please note that due to the nature of the work (journalism) there will be periods during which partipicants may choose to undertake independent research in order to write an article. This may involve traveling within Bolivia (as a tourist) and/or exploring La Paz (the city where we are based) to collect material for his pieces. We strive to team-up interns (in twos and threes, often with a local intern who knows the country) to undertake this research and all assignments are carried out on an entirely voluntary basis. By this we mean that interns are under no circumstances instructed to report on an event or go to a location on behalf of the publication, and have the option of limiting their involvement with the Bolivian Express to desk-based research. In short, an intern’s time at our offices (where we are able to provide constant supervision) are limited to attending meetings, writing articles (though they have the option of writing them in their own time, ie. at an internet cafe), designing the magazine, and carrying out administrative tasks.

To give a concrete example: interns often choose to visit nightclubs in La Paz during their personal time and then ask us whether we would like them to write a nightclub review based on their experiences. Of course, we accept this request and publish their reviews, without even instructing them to visit a nightclub.

To give another concrete example: some interns choose to travel within the country on weekends, or request time away for the office to do tourism. They later write articles about their experiences (some of which are published) yet they are never instructed to undertake any adventures (including extreme sports).

Point 2 – Workplace insurance

Workplace insurance is taken out for interns, and covers them during their time at our offices or when they are carrying out their duties elsewhere on our instruction. However, all interns are required to take out comprehensive travel insurance before arriving, covering them for any eventuality during their independent travel and research.

Point 3 –  Risk Assessment

In terms of risk assessment, we require interns to fully research their proposed travel plans in the relevant official (ie, foreign office travel advice) and unofficial literature (ie, travel guides), and we give them a safety briefing before their departure with additional contextual information acquired through personal experience.

Furthermore, we require interns to carry a Bolivian mobile phone with them at all times so they can inform us immediately in the case of an emergency. They are also required to carry cards with their address and key telephone numbers (provided) next to their press passes, which they must wear whilst undertaking any work for the magazine.

We also instruct interns to under no circumstances take an unlicensed form of transport, and provide them with the means to avoid this situation (mobile phones and information on how to order a taxi). However, many still choose to take unlicensed taxis, so we are unable to accept any form of liability in these cases.

Point 4 – First Aid – re: “an adequate number of first aiders”

We are a small magazine. There are 2-4 full time staff on the ground at any given time. While some of us have knowledge of First Aid practices not all of us have undertaken formal training in this area. To mitigate the risks associated with this situation, we are registered with a 24hr clinic five minutes away from the interns’ residence and the offices of the Bolivian Express, and have successfully (and immediately) been able to refer those suffering from food poisoning and altitude sickness (very common in Bolivia) to this clinic, accompanied by another person. We also ask interns who have received training in this area to inform us upon arrival in order to inform everyone taking part in the programme in case any basic first-aid tasks are required.

Guest post: Why I escaped The Times’ paywall

In a guest post, blogger Tim Kevan explains why he resigned from The Times over the paywall

Back in early 2007 I had been practising as a lawyer for some nine years. But I’d always dreamt of living by the sea and the surf and maybe even writing a novel. I just couldn’t quite see how it could be done.When I finally sat down to write a legal thriller what popped out instead was a legal comedy about a fictional young barrister doing pupillage.

I called him BabyBarista which was a play on words based on his first impression being that his coffee-making skills were probably as important to that year as any forensic legal abilities he may have. I wrote it as a blog and was hopeful it might raise a few smiles but in my wildest dreams I hadn’t imagined quite the extraordinary set of circumstances which then unfolded with The Times offering to host the blog and Bloomsbury Publishing of Harry Potter fame offering to make it into a book.

Since then the first book came out last August and was originally called BabyBarista and the Art of War. It is being re-issued in August under the new title Law and Disorder and the sequel is due out next May.

I was also continuing to publish my blog on The Times until May this year when it became clear that even blogs were going to go behind their new paywall. Continue reading

Magazines and digital: a report from the PTC Academies and Industry Forum

Suzanne Kavanagh reports on key insights and highlights from the Periodicals Training Council (PTC) Academies and Industry Forum, at Bauer Media’s central London office.

Editorial is at the heart of management at Bauer, said the company’s CEO, Paul Keenan, who explained how they work across media and events for brands and are embracing digital.

Keenan provided several insights into the industry and Bauer’s business – helpful information for anyone applying to get into the industry: Continue reading

Quackwatch sued by Doctor’s Data

A familiar story. Here’s the rundown from The Quackometer:

“Stephen Barrett [of Quackwatch] has been very critical of [Doctor’s Data] and has written that the diagnostic health tests it provides are used to defraud patients. One test in particular stood out for his criticism where patients are given a “provoking agent” that flushes out heavy metals into the urine. A urine test is then analysed by DDI and the concentration of heavy metals is compared with standards. Except the standards used are for patients who have not had the provoking agent. The levels of metals are going to be much higher than normal and this ‘elevated result’ is then used to sell expensive and unnecessary treatments.”

Sounds like a valid subject to investigate. Then:
Continue reading

When Open Public Data Isn’t…?

This year was always going to be an exciting year for open data. The launch of data.gov.uk towards the end of last year, along with commitments from both sides of the political divide before the election that are continuing to be acted upon now means data is starting to be opened up -scruffily at first, but that’s okay – and commercial enterprises are maybe starting to get interested too…

…which was always the plan…

…but how is it starting to play out?

The story so far…

A couple of weeks ago, the first meeting of the Public Data Transparency Board was convened, which discussed – and opened up for further discussion, a set of draft public data principles. (Papers relating to the meeting can be found here.)

In a letter to the responsible Minister prior to the meeting (commentable extracts here), Professor Nigel Shadbolt suggested that:

4. … The economic analysis, and the views we regularly hear from the business community themselves, are unequivocal: data must be released for free re-use so that the private sector can add new value and develop innovative new business services from government information. …

8. Transparency principles need to be extended to those who operate public services on a franchised, regulated or subsidised basis. If the state is controlling a service to the public or is franchising or regulating its delivery the data about that activity should be treated as public data and made available. …

11. We need to support the development of licences and supporting policies to ensure that data released by all public bodies can be freely re-used and is interoperable with the internationally recognised Creative Commons model. …

12. A key Government objective is to realise significant economic benefits by enabling businesses and non-profit organisations to build innovative applications and websites using public data. …

The business imperative is further reinforced by the second of three reasons given by the Open Government Data tracking project in Why Open Government Data?:

Releasing social and commercial value. In a digital age, data is a key resource for social and commercial activities. Everything from finding your local post office to building a search engine requires access to data much of which is created or held by government. By opening up data, government can help drive the creation of innovative business and services that deliver social and commercial value.

So how has business been getting involved? As several local councils start to pick up a request contained in a letter from the Prime Minister published at the end of May that they open up their financial data, Chris Taggart/@countculture, developer of OpenlyLocal posted a piece on The open spending data that isn’t… this is not good in which he described how apparently privileged access to financial data from several councils was being used to drive Spikes Cavell’s SpotlightOnSpend website (for a related open equivalent, see Adrian Short’s Armchair Auditor). Downstream use of the data was hampered by a “personal use only” license, and a CAPTCHA that requires a human in the loop in order to access the data. The Public Sector Transparency Board promptly responded to Chris’ post (Work on Local Spending Data), quoting the principle that:

“Public data will be released under the same open licence which enables free reuse, including commercial reuse – all data should be under the same easy to understand licence. Data released under the Freedom of Information Act or the new Right to Data should be automatically released under that licence.”

and further commenting: “We have already reminded those involved of this principle and the existing availability of the ‘data.gov.uk’ licence which meets its criteria, and we understand that urgent measures are already taking place to rectify the problems identified by Chris.”

Spikes Cavell chief executive Luke Spikes responded via an interview with Information Age, (SpotlightOnSpend reacts to open criticism):

[SpotlightOnSpend] is first and foremost a spend analysis software and consultancy supplier, and that it publishes data through SpotlightOnSpend as a free, optional and supplementary service for its local government customers. The hope is that this might help the company to win business, he explains, but it is not a money-spinner in itself.

“The contribution we’re making to transparency is less about what the purists would like to see, it’s simply putting the information out there in a form that is useful for the audience for which it is intended [i.e. citizens and small businesses]” he said. “But there are a few things we haven’t done right, and we’ll fix that.”

Following the criticism, Cavell says that SpotlightOnSpend will make the data available for download in its raw form. “That’s what we thought was the most sensible solution to overcoming this obstacle,” he told me.

Adrian Short, developer of the open Armchair Auditor, then picked up the baton in a comment to the Information Age article:

There is room for Spikes Cavell to develop their applications and I doubt that anyone has any objection to them offering their services to councils commercially just like thousands of other businesses. But they do not have a monopoly of ideas, talent and resources to build great applications with public spending data. Nor does anyone else.

The concerns that @CountCulture raised were not that Spikes Cavell were trading with councils or trying to attract their business but that they are doing so in a way that precludes anyone else developing applications with this data. By legally and technically locking the data into the Spotlight on Spend platform, everyone else is excluded.

It’s understandable that most councils have no understanding of the culture, legalities or technicalities of open data. This is new territory for nearly all of them. Those councils that have put their data straight onto Spotlight on Spend, bypassing the part where it is made genuinely open — cannot be criticised for not complying with what to them must be a very unusual requirement. But that’s why @CountCulture and I and others want to be very clear that the end result of this process is having effective scrutiny of council finances through multiple websites and applications, not just Spotlight on Spend or any other single website or application. The way we get there is with open data.

And Chris Taggart’s response? (Update on the local spending data scandal… the empire strikes back):

Lest we forget, Spikes Cavell is not an agent for change here, not part of those pushing to open public data, but in fact has a business model which seems to be predicated on data being closed, and the maintenance of the existing procurement model which has served us so badly.

(For recommendations on how councils might publish financial data in an open way, see: Publishing itemised local authority expenditure – advice for comment (reiterated here: Open Government Data: Finances. The Office for National Statistics occasionally releases summary statistics (e.g. as republished in Openlocal: Local spending data in OpenlyLocal, and some thoughts on standards) but at only a coarse resolution. As to how much it might cost to do that, some are claiming Cost of publishing ‘£500 and above’ council expenditure prohibitive.)

From my own perspective, I would also add that should consultants like Spikes Cavell create derived data works from open public data, there should be some transparency in declaring how the derived work was created (see for example: So Where Do the Numbers in Government Reports Come From? and Data is not Binary).

Another example of how once open data is becoming “closed” behind a paywall comes from Paul Geraghty (“Closed Data Now” SOCITM does a “Times”):

If my memory serves me well the e-Gov Register (eGR) hosted by Brent has been showing every IT supplier sortable by product type, supplier, local government type and even on maps for about 6 or 7 years (some links below if you hurry up).

I am aware that there are problems with this data, in my own past employer I know that some of the data is out of date.

But it is there, it is useful and informative and it is OPEN to all, even SMEs like me researching on niche markets in local government.

The latest move by SOCITM (and presumably with the knowledge of the LGA and the IDeA) means all that data is going to go behind the SOCITM paywall.

And the response from Socitm, via a comment from Vicky Sargent:

First of all, I’d like to clear up some points of fact. No local authority or other public sector service provider that provides data to the Applications Register will have to pay for their subscription and for them, access to the data will be free, regardless of whether they subscribe to Socitm Insight (as 95% of local authorities do). Anyone who is employed in an organisation that is an Applications Register subscriber – f-o-c or paid, will be able to access the data.
Then there is who pays. Clearly an information service like this that adds value, has to cover the costs of development and delivery. Unlike government departments, LGA, IDeA and local councils, Socitm is not directly funded by the taxpayer, and needs to fund the services it delivers from money raised from fees, subscriptions, events and other services.
The business model we use for the Applications Register is that public bodies that contribute should not pay to use the service, but those that do not contribute pay in cash. Private sector bodies can only pay in cash.

Your article also suggests that Socitm’s support for the move towards open data is hypocritical, set against our business model for the Applications Register. I think this misunderstands the thinking behind ‘open data’, which is to get raw data out of government systems for transparency purposes, also so that it can be re-used. Socitm has been a long-term strong supporter of this.
The open data agenda explicitly acknowledges that ‘re-use’ includes adding value and selling on. If councils were to routinely publish the sort of data we will collect for the Applications Register, there would still be work to be done aggregating and manipulating and re-publishing the information to make it useful, and that is what we do, recovering our costs in the way described.

Adrian Short (can you see how it’s the same few players engaging in these debates?!;-) develops the “keep it free” argument in a further comment:

Your argument presupposes your conclusion, which is that Socitm is the best organisation to be managing/publishing the applications register. Because, as you correctly say, you don’t receive any direct funding from the taxpayer, you have to find other ways of paying for that work. Inevitably this means charging non-contributing users.

What you’re missing is that millions of pounds of public money is spent every year supporting businesses, helping to create new markets and generally oiling the parts of the economy that don’t easily oil themselves. That’s what BIS and the economic development departments of local authorities do. The public interest and private benefit aren’t easily distinguishable unless you contrive that private benefit for a small group to the exclusion of others. But as Paul rightly points out, the potential market for this information is enormous — essentially every business and individual that works for, supplies or wants to work for the public sector, from the individual IT worker to the massive global consultancies, manufacturers and software firms.

Currently it’s a small number of incumbent suppliers that benefit from this relatively inefficient market. Other businesses lose. Public sector buyers lose. The taxpayer loses.

Keeping this information free for everyone to use and enabling it to be used in future when combined with the enormous amount of data that will be released soon will be likely to produce economic benefits to the public through market efficiencies that outstrip its cost by several orders of magnitude. If Socitm can’t publish this data in the most useful, non-discriminatory way then it’s not the best organisation for the job. I can see no reason in principle or practice why it shouldn’t be fully funded by the taxpayer and free at the point of use for everyone. To do otherwise would be an extremely false economy.

(Note that “free vs. open” debates have also been played out in the open source software arena. Maybe it’s worth revisiting them…?)

The previously quouted comment from Vicky Sargent also contains what might be described as an example case study:

This brings me to Better Connected, the annual survey of council websites carried out by Socitm. You say:
Just about every council in the UK has little option but to pay SOCITM hundreds of pounds annually to join their club to find out the exact details of how their website is being ranked.The data for Better connected only exists because Socitm has devised a methodology for evaluating websites, pays for a team of reviewers collect the data each year, and then analyses and publishes the results. No one has to subscribe, they choose to do so because the information is valuable to them.
Information about how we do the evaluation and ranking is freely available on our website, in our press releases and in our free-to-join Website Usage and Improvement community. The 2010 headline results for all councils are published on socitm.net as open data under a creative commons licence and are linked from data.gov.uk.
If the Better connected report has become a ‘must read’, that is because the investment Socitm has made in the product has led to it being a more cost-effective investment for councils than alternative sources of advice on improving their website. Many users have told us Better connected (cover price £415 for non-subscribers or free as part of the Socitm Insight subscription that starts at £670 pa for a small district council) is worth many days’ consultancy, even when that consultancy is purchased from lower cost SME providers.

As these examples show, the license under which data is originally released can have significant consequences on its downstream use and commercialisation. The open source software community has know this for years, of course, which is why organisations like GNU have two different licenses – GPL, which keeps software open by tainting other software that includes GPL libraries, and LGPL, which allows libraries to be used in closed/proprietary code. There is a good argument that by combining data from different open sources in a particular way valuable results may be created, but it should also be recognised that work may be expended doing this and a financial return may need to be generated (so maybe companies shouldn’t have to open up their aggregated datasets?) Just how we balance commercial exploitation with ongoing openness and access to raw public data is yet to be seen.

(The academic research area – which also has it’s own open data movement (e.g. Panton Principles) – also suggests a different sort of tension arising from the “potential value” of a data set or aggregated data set. For example, research groups analysing data in one particular way may be loathe to release to others because they want to analyse it in another, value creating way at a later date.)

Getting the licensing right is particularly important if councils become obliged to use third party services to publish their data. For example, the grand vision of the Public Sector Transparency Board identified in this paragraph in Shadbolt’s letter to Maude states:

13. We must promote and support the development and application of open, linked data standards for public data, including the development of appropriate skills in the public services. …

But as a recent report, again from Chris Taggart, on Publishing Local Open Data – Important Lessons from the Open Election Data project suggests, there are certain challenges associated with web related development in local authorities, and in particular a significant lack of experience and expertise in dealing with Linked Data (which is not surprising – it is a relatively new, and so far arcane) technology. Here are the first four lessons, for example:

– There is a lack of ‘corporate’ awareness/understanding of open data issues, and this will inhibit take up of open, linked data publishing unless it is addressed
– There is a lack of even basic web skills at some councils
– Many councils lack web publishing resources, never mind the resources to implement open, linked data publishing
– The understanding of even the basics of linked data and the steps to publishing public data in this way is very, very limited

What this suggests to me is that it is likely that in the short term at least, the capability for publishing Linked Data will reside in specialist third party companies, possibly one of only a few companies. As Paul Geraghty discovers from the eGovernment Register in If #localgovweb supplier says “RDF WTF?” Sack em #opendata #spending:

[I]t seems to me that of 450 or so local government organisations, 357 are listed as having a “Financials” supplier **.

There are only 18 suppliers listed, and of those there are 6 Big Ones.

Between them the 6 Big Ones supply “Financials” to 326 Councils.

Don’t you think that the first one of those 6 Big Ones who natively supports LOD [Linked Open Data] as an export option (or agrees to within, say, 8 months) really ought to be favoured when bidding for new business?

Lets go further, lets say that it should be mandated that all new contracts with “Financials” suppliers include an LOD clause.

Perhaps Mr Pickles could dispatch someone to have a chat with one or two of these suppliers, or that he should have someone check that future contracts for Financial products being sold to Local Government all contain the necessary wording to make this happen?

So instead of trying to train and cajole 450 councils to FTP assorted CSV files into localdata.gov.uk (FFS) all the way through to grokking RDF, namespaces, LOD et al – why does the government not get on and make a strategy to bully and coerce 6 suppliers instead – and potentially get 326 councils teed up to produce useful LOD a bit sharpish?

Another technology option is for councils to publish their own linked data to a commercially hosted datastore. At the moment, the two companies I know of that offer “datastore” services for publishing Linked Data, at scale, are Talis, and the Stationery Office (in partnership with Garlik). It is, of course, open knowledge that one Professor Nigel Shadbolt is a director of Garlik Limited.

An introduction to data scraping with Scraperwiki

Last week I spent a day playing with the screen scraping website Scraperwiki with a class of MA Online Journalism students and a local blogger or two, led by Scraperwiki’s own Anna Powell-Smith. I thought I might take the opportunity to try to explain what screen scraping is through the functionality of Scraperwiki, in journalistic terms.

It’s pretty good.
Continue reading

Don't stop us digging into public spending data

A disturbing discovery by Chris Taggart last week: a number of councils in the UK are handing over their ‘open’ data to a company which only allows it to be downloaded for “personal” use.

As Chris himself points out, this runs completely against the spirit of the push to release public data in a number of ways:

  • Data cannot be used for “commercial gain”. This includes publishers wanting to present the information in ways that make most sense to the reader, and startups wanting to find innovative ways to involve people in their local area. Oh, and that whole ‘Big Society‘ stuff.
  • The way the sites are built means you couldn’t scrape this information with a computer anyway
  • It’s only a part of the data. “Download the data from SpotlightOnSpend and it’s rather different from the published data [on the Windsor & Maidenhead site]. Different in that it is missing core data that is in W&M published data (e.g. categories), and that includes data that isn’t in the published data (e.g. data from 2008).”

It’s a worrying path. As Chris sums it up: ” Councils hand over all their valuable financial data to a company which aggregates for its own purposes, and, er, doesn’t open up the data, shooting down all those goals of mashing up the data, using the community to analyse and undermining much of the good work that’s been done.”

The Transparency Board quickly issued a statement about this issue saying that “urgent” measures are taking place to rectify the problem.

And Spikes Cavell, who make the software, responded in Information Age, pointing out that “it is first and foremost a spend analysis software and consultancy supplier, and that it publishes data through SpotlightOnSpend as a free, optional and supplementary service for its local government customers. The hope is that this might help the company to win business, he explains, but it is not a money-spinner in itself.”

They are now promising to make the data available for download in its “raw form”, although it’s not clear what that will be. Adrian Short’s comment to the piece is worth reading.

Nevertheless, this is an issue that anyone interested in holding power to account should keep a close eye on. And to that aim, Chris has started an investigation on Help Me Investigate to find out how and why councils are giving access to their spending data. Please join it and help here.

(Comment or email me on paul at helpmeinvestigate.com if you want an invitation.)

77,000 pageviews and multimedia archive journalism (MA Online Journalism multimedia projects pt4)

(Read part 1 here; part 2 here and part 3 here)

The ‘breadth portfolio’ was only worth 20% of the Multimedia Journalism module, and was largely intended to be exploratory, but Alex Gamela used it to produce work that most journalists would be proud of.

Firstly, he worked with maps and forms to cover the Madeira Island mudslides:

“When on the 20th of February a storm hit Madeira Island, causing mudslides and floods, the silence on most news websites, radios and TV stations was deafening. But on Twitter there were accounts from local people about what was going on, and, above all, they had videos. The event was being tagged as #tempmad, so it was easy to follow all the developments, but the information seemed to be too scattered to get a real picture of what was going on in the island, and since there was no one organizing the information available, I decided to create a map on Google[ii], to place videos, pictures and other relevant information.

“It got 10,000 views in the first hours and reached 30,000 in just two days. One month later, it has the impressive number of 77 thousand visits.”

Not bad, then.

Secondly, Alex experimented with data visualisation to look at newspaper brand values and the online traffic of Portuguese news websites.

“My goal was to understand the relative and proportional position of each one, regarding visits, page views, and how those two values relate to each other. The data I got also has portals, specialized websites, and entertainment magazines so it has a broad range of themes (all charts are available live here – http://is.gd/aZLXs)”

And finally, he produced a beautiful Flash interactive on Moseley Road Baths (which he talks about here).

All of which was produced and submitted within the first six weeks of the Multimedia Journalism module.

The other 80%: multimedia archive journalism

Alex was particularly interested in archive journalism and using multimedia to bring archives to life. As a way of exploring this he produced the Paranoia Timeline, a website exploring “all the events that caused some type of social hysteria throughout the world in the last 20 years.

“Some of the situations presented here were real dangers, others not really. But all caused disturbances in our daily lives … Why does that happen? Why are we caught in these bursts of information, sometimes based on speculative data and other times borne out of the imagination of few and fed by the beliefs of many?”

The site – which is an ongoing project in its earliest stages – combines video, visualisation, a Dipity timeline, mapping and the results of some fascinating data and archive journalism. Alex explains:

“The swine flu data came from Wolfram-Alpha[vi] that generated a rather reliable (after cross checking with other official websites) amount of data, with the number of cases and deaths per country. I had to make an option about which would be highlighted, but discrepancies in the logical amount of cases between countries made me go just for the death numbers. The conclusion that I got from the map is that swine flu was either more serious or reported in the developed countries. Traditionally considered Third World countries do not have many reports, which reflect the lack of structures to deal with the problem or how overhyped it was in the Western world. But France on its own had almost 3 million cases reported against 57 thousand in the United States, which led me to verify closely other sources. It seems Wolfram Alpha had the number wrong, there were only about 5000 reports, which proves that outliers in data are either new stories or just input errors.

“For the credit crunch[vii], I researched the FDIC – Federal Deposit Insurance Corporation[viii] database. They have a considerable amount of statistical data available for download. My idea was to chart the evolution of loans in the United States in the last years, and the main idea was that overall loans slowed down since 2009 but individual credits rose, meaning an increase in personal debt to cope with overall difficulties caused by the crunch.I selected the items that seemed more relevant and went for a simple line chart. My purpose was served.”

“Though the current result falls short of my initial goals,” says Alex, “it is a prototype for a more involving experience, and I consider it to be a work in construction. What I’ll be defending here is a concept with a few examples using interactive tools, but I realize this is just a small sample of what it can really be: an immersive, ongoing project, with more interactive features, providing a journalistic approach to issues highly debated and prone to partisanship, many of them used by religious and political groups to spin their own ideologies to the general audience. The purpose is to create context.”

Alex is currently back in Portugal as he completes the final MA by Production part of his Masters. You might want to hire him, or Caroline, Dan, Ruihua, Chiara, Natalie or Andy.