Another Word For It Patrick Durusau on Topic Maps and Semantic Diversity

December 18, 2015

Buzzfeed uses R for Data Journalism

Filed under: Journalism,News,R,Reporting — Patrick Durusau @ 11:36 am

Buzzfeed uses R for Data Journalism by David Smith.

From the post:

Buzzfeed isn't just listicles and cat videos these days. Science journalist Peter Aldhous recently joined Buzzfeed's editorial team, after stints at Nature, Science and New Scientist magazines. He brings with him his data journalism expertise and R programming skills to tell compelling stories with data on the site. His stories, like this one on the rates of terrorism incidents in the USA, often include animated maps or interactive charts created with R. 

Data journalists and would be data journalists should be following the use of R and Python at Buzzfeed.

You don’t have to read Buzzfeed (I have difficulty with its concept of “news”), as David points out a way to follow all the Buzzfeed projects that make it to GitHub.

See David’s post for other great links.

Enjoy!

December 16, 2015

Avoiding the Trap of Shallow Narratives

Filed under: Journalism,Narrative,News,Reporting — Patrick Durusau @ 8:42 pm

Avoiding the Trap of Shallow Narratives by Tiff Fehr.

From the post:


When we elevate immediate reactions to the same level as more measured narratives, we spring a trap on ourselves and our readers. I believe by the end of 2016, we will know if a “trap” is the right description. 2016 is going to be turbulent for news and news-reading audiences, which will add to the temptation to chase traffic via social-focused follow-on stories, and perhaps more of clickbait’s “leftover rehash.” Maybe we’ll even tweak them so they’re not “a potential letdown,” too: “Nine Good Things in the SCOTUS Brawl at the State of the Union.”

A great read on a very serious problem, if your goal is to deliver measured narratives of current events to readers.

Shallow narratives are not a problem if your goals are:

  • First, even if wrong, is better than being second
  • Headlines are judged by “click-through” rates
  • SEO drives the vocabulary of stories

This isn’t a new issue. Before social media, broadcast news was too short to present any measured narrative. It could signal events that needed measured narrative but it wasn’t capable of delivering it.

No one watched the CBS Evening News with Walter Cronkite to see a measured narrative about the Vietnam War. For that you consulted Foreign Affairs or any number of other history/policy sources.

That’s not a dig at broadcast journalism in general or CBS/Cronkite in particular. Each medium has its limits and Cronkite knew those limits as well as anyone. He would have NOT warned anyone off who was seeking “measured narrative” to supplement his reports.

The article I mentioned earlier about affective computing, We Know How You Feel from the New Yorker, qualifies as a measured narrative.

As an alternative, consider the shallow narrative: Mistrial in Freddie Gray Death. Testimony started December 2nd and the entire story is compressed into 1,564 words? Really?

Would anyone consider that to be a “measured narrative?” Well, other than its authors and colleagues who might fear a similar evaluation of their work?

You can avoid the trap of shallow narratives but that will depend upon the forum you choose for your content. Pick something like CNN and there isn’t anything but shallow narrative. Or at least that is the experience to date.

Your choice of forum has a much to do with avoiding shallow narrative as any other factor.

Choose wisely.

December 14, 2015

Essential 2016 Trends Overload Here!

Filed under: Forecasting,Journalism,News,Reporting — Patrick Durusau @ 7:55 pm

OUTLOOK ’16 /// Essential trends for 2016 by Ezra Eeman.

From the post:

The world is changing. In small iterations and disruptive shifts. New platforms emerge, new technologies shape our behaviour. The need for new business models and new talent was never more urgent. In order to stay ahead newsroom leaders and journalists have to look forward and understand the emerging trends that are/will be changing our daily lives.

OUTLOOK ‘16/// is a broad & growing collection of trend reports about media, technology and society. The selection is handmade in collaboration with VRT Start-Up. The order is random.

There are 17 trend reports as I write this post and no doubt more will be added as 2016 inches ever closer.


This is a growing collection. New relevant reports will be added when they are released. A Dutch version of this collection can be found here.

Want to suggest a great trend report for this collection? Mail us at journalismtools.mailbox@gmail.com

Hat tip to Journalism Tools for the collecting, which saves all of us the time of mining for 2016 trend reports.

Are you going to read all seventeen reports? Or whatever the ultimate number of trend reports?

How would you curate all seventeen+ reports to enable others to quickly survey the results and dip in and out of them?

PS: The “4 min read” is deceptive. You can scan the titles of all 17 trend reports in 4 minutes but be woefully short of reading time for all of them.

No Sign of Terrorist Attack = Conflict in Government Priorities

Filed under: Journalism,News,Newspeak,Reporting — Patrick Durusau @ 3:09 pm

Egypt Says Investigators Found No Sign Of ‘Terrorist Act’ In Russian Plane Crash by Eyder Peralta.

Despite early conclusions by the Russians, the United States, Britain, and claims by the Islamic State, the Egyptian government has concluded there is no sign of a terrorist attack in the downing of a Russian passenger plane in Egypt last October.

Reporters and citizens alike should view claims of “terrorist” and “not a terrorist” attack with heavy additions of salt.

In this particular case, Egypt wants to avoid further damage to its revenue from tourism, which is reported down by 10% over last year.

U.S. intelligence and law enforcement agencies are desparately trying to find a terrorist connection for the shooters in San Bernardino. See also: Everything we know about the San Bernardino terror attack investigation so far.

From the second Los Angeles Times story:

Why their plot wasn’t detected

Farook and Malik were unknown to law enforcement until the day of the shooting, but the reason for that isn’t yet clear.

The FBI is focusing on how they missed the couple’s secret radicalization and Farook’s apparent comments to an associate as early as 2011 that he was considering a terrorist attack.

Two people out of a current population of 322,332,500 (as of 14:58 EST today) not being known to law enforcement doesn’t surprise me.

Does it surprise you?

As far as threats of a terrorist attack years ago, if they arrested everyone who shouts “kill the umpire/referee, etc.,” there would not be anyone left to stand as guards in the prisons.

The reason for treating the deaths of co-workers at a holiday party as an act of terrorism is that it furthers the budget agendas of law enforcement and intelligence communities.

Imagine the surveillance that would be required to gather and be aware of a random statement from four years ago from otherwise unremarkable individuals.

Your system would have to connect a statement to a co-worker to the purchases of the weapons, ammunition and other supplies and then ring the alert bell to clue officers in on a pending threat.

That’s possible in retrospect but to prevent random acts your system would have to do all those connections in the absence of any reason to focus on these individuals in particular.

You know the old saying, when they criminalize guns, only criminals will have guns?

Same is true for privacy, when they criminalize privacy, only criminals will have privacy.

PS: Remember terrorism is a label used to question the loyalty, judgement of others and/or for furthering other agendas. You could substitute “belch” where you see it and still have 99% of the informative content of a message.

December 10, 2015

The Quartz guide to bad data

Filed under: Journalism,News,Reporting — Patrick Durusau @ 11:50 am

The Quartz guide to bad data

From the webpage:

An exhaustive reference to problems seen in real-world data along with suggestions on how to resolve them.

As a reporter your world is full of data. And those data are full of problems. This guide presents thorough descriptions and possible solutions to many of the kinds of problems that you will encounter when working with data.

Most of these problems can be solved. Some of them can’t be solved and that means you should not use the data. Others can’t be solved, but with precautions you can continue using the data. In order to allow for these ambiguities, this guide is organized by who is best equipped to solve the problem: you, your source, an expert, etc. In the description of each problem you may also find suggestions for what to do if that person can’t help you.

You can not possibly review every dataset you encounter with for all of these problems. If you try to do that you will never get anything published. However, by familiarizing yourself with the kinds of issues you are likely to encounter you will have a better chance of identifying an issue before it causes you to make a mistake.

If you have questions about this guide please email Chris. Good luck!

I hesitate at the word exhaustive for all but the most trivial of collections.

Saying “this is a guide to the most common bad data problems encountered by journalists” comes closer to the mark.

It makes a great checklist to develop your own habit of routine data checks that you apply to all data, even data from trusted sources. (Perhaps even more so from trusted sources.)

Enjoy!

December 4, 2015

Witness (activists/change advocates, please read)

Filed under: Journalism,News,Reporting — Patrick Durusau @ 5:15 pm

Witness

From the about page:

WITNESS trains and supports activists and citizens around the world to use video safely, ethically, and effectively to expose human rights abuse and fight for human rights change.

The majority of the world’s population now has a camera in their pocket. People everywhere are turning to video to document and tell stories of abuse. But all too often, they are not filming safely or effectively, and their videos don’t make a difference.

We identify critical situations and teach those affected by them the basics of video production, safe and ethical filming techniques, and advocacy strategies.

A wealth of training materials and videos for anyone who wants to make a difference with their cellphone or video camera.

TV images of the police using water cannon and setting dogs on young people and children in Birmingham, Alabama, were a watershed event for the civil rights movement in the United States.

But that isn’t to promise the technique works every time. Scenes of far worse abuse are common from the Gaza strip and world governments continue to look away.

But you won’t know if the video advocacy will work for you or not unless you try. The Witness site will enable you to do just that.

Many of the techniques described are equally useful for journalists, researchers, etc.

I first saw this at FirstDraftNews as: The Witness Resource Library.

PS: Since I mentioned the Birmingham Children’s Crusade (1963), I should also point out that Malcolm X disagreed with the strategy of sending children into harm’s way. As do I. There are more appropriate ways to confront those of practice violence than sacrificing children for photo ops.

December 3, 2015

How-To Detect Fake Images in 20 Seconds

Filed under: Journalism,News,Reporting,Verification — Patrick Durusau @ 11:24 am

The video has a good “beat” to it. 😉

Entire video is 42 seconds.

Enjoy!

Verification around the world: From Kenya to Hong Kong

Filed under: Journalism,News,Reporting,Verification — Patrick Durusau @ 10:47 am

Verification around the world: From Kenya to Hong Kong by Tom Trewinnard.

From the post:

Misinformation around news on the social web is a truly global issue and around the world there are many groups and projects rising to the verification challenge. Here’s a by-no-means-exhaustive run down of a few projects we at Meedan have been watching recently.

Have you seen any great international verification efforts that we’ve missed? Share them with us in the comments here or at @FirstDraftNews on Twitter.

Seven very good examples of verification from a variety of news outlets.

You don’t want to include false information in your topic maps.

How do you verify information for inclusion?

When you debunk rumors or uncover false information, how do you share it with others?

December 2, 2015

Twitter Journalism Tips

Filed under: Journalism,News,Reporting,Tweets,Twitter — Patrick Durusau @ 1:49 pm

Twitter Journalism Tips from FirstDraftNews.

Five videos on effective use of Twitter for journalism.

The videos are:

How To Use Twitter Lists For Journalism 2:37

Why I Love Twitter Lists – Sue Llewellyn 1:48

Journalist Guide: How To Use Tweekdeck 1:14

Journalist Tweetdeck Tips – Reuters, George Sargent 1:12

Searching For Geolocated Posts On Twitter 1:26

The times shown are minutes followed by seconds.

Labeled for journalism but anyone searching Twitter, librarians, authors, researchers, even “fans” (shudder), will find useful information in these videos.

If you don’t know FirstDraftNews, you need to get acquainted.

December 1, 2015

Are you a debunker or fact-checker? (take the survey, it’s important)

Filed under: Journalism,News,Reporting,Social Media — Patrick Durusau @ 7:17 pm

Are you a debunker or fact-checker? Take this survey to help us understand the issue by Craig Silverman.

From the post:

Major news events like the Paris attacks quickly give birth to false rumors, hoaxes and viral misinformation. But there is a small, growing number of people and organizations who seek to quickly identify, verify or debunk and spread the truth about such misinformation that arises during major news events. As much as possible, they want to stop a false rumor before it spreads.

These real-time debunkers, some of which First Draft has profiled recently, face many challenges. But by sharing such challenges and possible solutions, it is possible to find collective answers to the problem of false information.

The survey below intends to gather tips and tactics from those doing this work in an attempt to identify best practices to be shared.

If you are engaged in this type of work — or have experimented with it in the past — we ask that you please take a moment to complete this quick survey and share your advice.

I don’t know if it is possible for debunking or fact-checking to run faster than rumor and falsehood but that should not keep us from trying.

Definitely take the survey!

November 28, 2015

The First Draft Toolbox for newsgathering and verification

Filed under: Journalism,News,Reporting — Patrick Durusau @ 9:22 pm

If you are not Donald Trump or some other form of a pathological liar, then you will enjoy: The First Draft Toolbox for newsgathering and verification by Alastair Reid.

From the post:

Welcome to the First Draft Toolbox, a list of tools and sites recommended by the First Draft Coalition to help in social newsgathering, verification and more.

We will be updating the page regularly with new tools as well as more detailed explainers and guides of those listed here already. If you have any suggestions of something we may have missed or are launching a tool you think should be featured here, please let us know by emailing our editor Alastair Reid.

You can also get email alerts for when we update the page using ChangeDetection or other available tools.

So many options can be overwhelming though, and putting them into practice can be daunting when just starting out. The best advice has always been to experiment with everything but find the tools that work for you, and keep up with thought leaders and case studies to see what the experts use and how they use them.

By rough count I make it thirty-eight separate resources for newsgathering and verification. The big categories are: Social newsgathering and search tools, Location checking tools, Source verification, Image verification, YouTube Data Viewer and, Translation.

An impressive collection, several new to me and more than you will probably use at any one time. Try the most needed ones first and then branch out. Over time you will develop favorites and skill at using them.

The one omission that surprised me was Alastair failing to mention Snopes.com.

Snopes.com is one of the premier debunking sites on the WWW. For example:

Undercover Parcel Service No, UPS isn’t smuggling refugees into the United States in the dead of night.

Cetacean Harvestation No, cranberry farmers aren’t netting and canning dolphins during the harvest season.

Does that help explain Donald Trump’s standings in the polls?

Ask not only whether statements are “true,” but also what the speaker has to gain from giving them to you?

November 26, 2015

What Should the Media Do When Donald Trump Blatantly Lies? [Try Not Reporting Lies]

Filed under: Journalism,News,Reporting — Patrick Durusau @ 8:07 pm

What Should the Media Do When Donald Trump Blatantly Lies? by Matthew Ingram.

From the post:

Political speech is a unique animal, especially during election season. It often mixes hyperbole with flowery language and aggressive rhetoric designed to inflame a particular passion. But Republican presidential candidate Donald Trump is arguably in a category unto himself. More than almost any other 2016 candidate, he is prone to telling flat-out lies, making up facts, and distorting the truth to a prodigious extent.

This kind of behavior creates a tricky problem for the press. How should media companies deal with Trump and his falsehoods? If he were just a joke candidate without a hope of ever being the Republican nominee, it would be easy enough to ignore him. But he appears to stand a better than even chance of getting the nomination — he has been leading in the polls for months.

If media outlets attack Trump’s lying directly, they run the risk of being accused of bias by his supporters and Republicans in general. In fact, that kind of reaction is already occurring in response to a New York Times editorial that accused the billionaire businessman of playing fast and loose with the truth on a number of issues, including whether Muslims in New Jersey cheered the Sept. 11, 2001 terrorist attacks.

Part of the problem is that Trump and his candidacy are to some extent a creation of the mainstream media. At the very least, the two have developed a disturbingly co-dependent relationship.

As disturbing as the article in on media coverage of lies by Donald Trump, the crux of the dilemma was put this way:

since the U.S. news media is based on the commercial model—and more eyeballs on the page or the screen is good for business—the networks love it when someone like Donald Trump says outrageous stuff. Fact-checking rains on the parade of that revenue model.

Perhaps news rooms need a new version of First they came for:

First Trump lied about the refugees, and I reported it—
Because I was not a refugee.

Then Trump lied about blacks, and I reported it—
Because I was not black.

Then Trump lied about Jews, and I reported it—
Because I was not a Jew.

Donald Trump lied his way into the Whitehouse, and I made it possible-
Because fact checking conflicted with the bottom-line.

When I think about journalists who risk their lives reporting on drug cartels and violent governments, I wonder what they must think of the moral cowardice of political coverage in the United States?

November 21, 2015

Leaking Classified Information

Filed under: Cybersecurity,Journalism,News,Reporting — Patrick Durusau @ 8:57 pm

I saw a tweet recently extolling the number of classified documents that could have obtained.

Not obtaining and/or leaking classified documents of any government denies the public information it can use.

Two suggestions:

If you can obtain classified information, do.

If you have classified information, leak it in its entirety.

Before some ambitious assistant US attorney decides I am advocating illegal activity, recall that some leaks of classified information are in fact authorized by the executive branch of the United States government. Read All Leaks Are Illegal, but Some Leaks Are More Illegal Than Others by Conor Friedersdorf for some example cases.

Classification is used to conceal embarrassing information or failures. No government has a right to conceal embarrassing information or failures.

Agonizing over what to leak creates power for those with leaked information from a government. Do you see yourself as that petty and vain?

Just leak it. Let the chips fall where they may.

The history of leaking is on the side of no harm to anyone.

Start with the Pentagon Papers (U.S. Archives), Watergate at 40, Public Library of US Diplomacy, which also includes Cablegate, the Kissinger cables and Carter cables parts 1 and 2, Afghan War Diaries, the Snowden leaks and count the bodies.

So far, I’ve got nothing. Zero. The empty set.

Over forty years of leaking and no bodies. If there was even one, it would be front and center at every leak story.

Doesn’t that tell you something about the truthfulness of government objections to leaks?

November 20, 2015

Four free online plagiarism checkers

Filed under: Journalism,News,Plagiarism,Reporting — Patrick Durusau @ 5:32 pm

Four free online plagiarism checkers

From the post:

“Detecting duplicate content online has become so easy that spot-the-plagiarist is almost a party game,” former IJNet editor Nicole Martinelli wrote in 2012. “It’s no joke, however, for news organizations who discover they have published copycat content.”

When IJNet first ran Martinelli’s post, “Five free online plagiarism checkers,” two prominent U.S. journalists had recently been caught in the act: Fareed Zakaria and Jonah Lehrer.

Following acknowledgement that he had plagiarized sections of an article about gun control, Time and CNN suspended Zakaria. Lehrer first came under scrutiny for “self-plagiarism” at The New Yorker. Later, a journalist revealed Lehrer also fabricated or changed quotes attributed to Bob Dylan in his book, “Imagine.”

To date, Martinelli’s list of free plagiarism checkers has been one of IJNet’s most popular articles across all languages. It’s clear readers want to avoid the pitfalls of plagiarism, so we’ve updated the post with four of the best free online plagiarism checkers available to anyone, revised for 2015:

Great resource for checking your content and that of others for plagiarism.

The one caveat I offer is to not limit the use of text similarity software solely to plagiarism.

Text similarity can be a test for finding content that you would not otherwise discover. Depends on how high you set the test for “similarity.”

And/or it may find content that is so similar, while not plagiarism (say multiple outlets writing from the same wire service) it isn’t worth the effort to read every story that repeats the same story with some minor edits.

Multiple stories but only one wire service source. In that sense, a “plagiarism” checker can enable you to skip duplicative content.

The post I quote above was published by the international journalist’s network (ijnet). Even if you aren’t a journalist, great source to follow for developing news technology.

November 19, 2015

In the realm of verification, context is king

Filed under: Journalism,News,Reporting — Patrick Durusau @ 4:45 pm

In the realm of verification, context is king by Fergus Bell.

From the post:

By thinking about the wider context around shared UGC you can often avoid a lengthy forensic verification process where it isn’t required. For publishers looking at how they tackle competition with platforms – it is easy. Context is where you can make a distinction through strong editorial work and storytelling.

Fergus has four quick tips that will help you fashion a context for user-generated content (UGC).

Content always has a context. If you don’t supply one, consumers will invent a context for your content. (They may anyway but you can at least take the first shot at it.)

It is interesting that user-generated content (UGC) isn’t held in high regard, yet news outlets parrot the latest rantings of elected officials and public figures as gospel.

When public statements are false, such as suggesting that Syrian refugees pose a danger of terrorism, why aren’t those statements simply ignored? Why mis-inform the public?

November 17, 2015

Lies, Damn Lies, and Viral Content [I Know a Windmill When I See One]

Filed under: Journalism,News,Reporting — Patrick Durusau @ 7:53 pm

Lies, Damn Lies, and Viral Content How News Websites Spread (and Debunk) Online Rumors, Unverified Claims and Misinformation by Craig Silverman.

From the executive summary:

News organizations are meant to play a critical role in the dissemination of quality, accurate information in society. This has become more challenging with the onslaught of hoaxes, misinformation, and other forms of inaccurate content that flow constantly over digital platforms.

Journalists today have an imperative—and an opportunity—to sift through the mass of content being created and shared in order to separate true from false, and to help the truth to spread.

Unfortunately, as this paper details, that isn’t the current reality of how news organizations cover unverified claims, online rumors, and viral content. Lies spread much farther than the truth, and news organizations play a powerful role in making this happen.

News websites dedicate far more time and resources to propagating questionable and often false claims than they do working to verify and/or debunk viral content and online rumors. Rather than acting as a source of accurate information, online media frequently promote misinformation in an attempt to drive traffic and social engagement.

The above conclusions are the result of several months spent gathering and analyzing quantitative and qualitative data about how news organizations cover unverified claims and work to debunk false online information. This included interviews with journalists and other practitioners, a review of relevant scientific literature, and the analysis of over 1,500 news articles about more than 100 online rumors that circulated in the online press between August and December of 2014.

Many of the trends and findings detailed in the paper reflect poorly on how online media behave. Journalists have always sought out emerging (and often unverified) news. They have always followed-on the reports of other news organizations. But today the bar for what is worth giving attention seems to be much lower. There are also widely used practices in online news that are misleading and confusing to the public. These practices reflect short-term thinking that ultimately fails to deliver the full value of a piece of emerging news.

Silverman writes a compelling account (at length, some 164 pages including endnotes) to prove:

News websites dedicate far more time and resources to propagating questionable and often false claims than they do working to verify and/or debunk viral content and online rumors. Rather than acting as a source of accurate information, online media frequently promote misinformation in an attempt to drive traffic and social engagement.

We have all had the experience of watching news reports where we know the “facts” and see reporters making absurd claims about our domain of expertise. But their words may be reaching millions and you can only complain to your significant other.

I fully understand Silverman’s desire to make news reporting better, just as I labor to impress upon standards editors the difference between a reference (is used in the standard itself) and further reading (as listed in a bibliography). That distinction seems particularly difficult for some reason.

The reason I mention windmills in my title is because Silverman offers this rationale for improving verification by news outlets:

Another point of progress for journalists includes prioritizing verification and some kind of value-add to rumors and claims before engaging in propagation. This, in many cases, requires an investment of minutes rather than hours, and it helps push a story forward. The practice will lead to debunking false claims before they take hold in the collective consciousness. It will lead to fewer misinformed readers. It will surface new and important information faster. Most importantly, it will be journalism.

The benefits are:

  1. Debunking false claims before they take hold in the collective consciousness
  2. Fewer misinformed readers
  3. Surface new and important information faster
  4. It will be journalism

Starting from the top: Debunking false claims before they take hold in the collective consciousness.

How does “debunking false claims” impact traffic and social engagement? If my news outlet doesn’t have the attention grabbing headline about an image of Mary in a cheese sandwich, don’t I lose that traffic? Do you seriously think that debunking stories have the audience share of fantastic claim stories?

I suppose if the debunking involved “proving” that the image of Mary was due to witchcraft, that might drive traffic but straight up debunking seems unlikely to do so.

The second benefit was Fewer misinformed readers.

I’m at a loss to say how “fewer misinformed readers” is going to benefit the news outlet? The consequences of being misinformed accrue to the reader and not to the news outlet. I suspect the average attention span is short enough that news outlets could take the other side tomorrow without readers being overly disturbed. They would just be misinformed in a different direction.

The benefit of Surface new and important information faster comes in third.

I can see that argument but that presumes that news outlets want to report “new and important information” in the first place. What Silverman successfully argues is the practice is to report news that drives traffic and social engagement. Being “new and important has only a tangential relationship to traffic and engagement.

You probably remember during the wall-to-wall reporting about Katrina or the earthquakes in Haiti the members of the news media interviewing each other. That was nearly negative content. Even rumors and lies would be have been better.

The final advantage Silverman cites is It will be journalism.

As I said, I’m not unsympathetic to Silverman but when was journalism ever concerned with not reporting questionable and false claims? During the American Revolution perhaps?, The Civil War?, WWI?, WWWII?, Korea?, Vietnam?, and the list goes on.

There have been “good” journalists (depending upon your point of view) and “bad” journalists (again depending on your point of view). Yet, journalism, just like theology, has survived being populated in part by scalawags, charlatans, and rogues.

What journalism needs is pro-active readers to rebel against superficial, inaccurate and misleading reporting. Voting with their feet will be far more effective than exhortations to do better.

November 15, 2015

On-Demand Data Journalism Training Site [Free Access Ends Nov. 30th]

Filed under: Journalism,News,Reporting — Patrick Durusau @ 10:18 pm

Investigative Reporters and Editors launches on-demand data journalism training site

From the post:

Want to become a data journalist? You’re going to need a lot of perseverance — as well as the right training.

To help make data journalism more accessible to all, Investigative Reporters and Editors (IRE) recently launched NICAR-Learn, an online platform of training videos that can be accessed from anywhere, at any time.

“NICAR-Learn is a place for journalists to demonstrate their best tricks and strategies for working with data and for others to learn from some of the best data journalists in the business,” IRE wrote in a statement.

Unlike many online training platforms, NICAR-Learn’s content won’t consist of hour-long webinars. Instead, NICAR-Learn will produce a library of short videos, often less than 10 minutes long, to train journalists on specific topics or techniques relating to data journalism.

The first NICAR-Learn videos come from data journalist MaryJo Webster, who has produced four tutorials that draw from her popular “Excel Magic” course. Users can request specific tutorials by submitting their ideas to IRE.

These videos will be available at no charge to non-IRE members through the end of November. Beginning in December, IRE will add more videos to NICAR-Learn and place them behind a paywall.

To learn more, visit NICAR-Learn’s “Getting Started” page.

I can’t say that I like “paywalls,” which I prefer to call “privilegewalls.”

Privilegewalls because that is exactly what paywalls are meant to be. To create a feeling of privilege among those who have access, to separate them from those who don’t.

And beyond a feeling of privilege, privilegewalls are meant to advantage insiders over those unfortunate enough to be outsiders. Whether those advantages are real or in the imagination of members I leave for you to debate.

Personally I think helping anyone interested to become a better journalist or data journalist will benefit everyone. Journalists, member of the public who read their publications, perhaps even the profession itself.

Here’s an example of where being a better “data journalist” would make a significant difference:

So far as I know no journalist, despite several Republican and Democratic presidential candidate debates has every asked how they propose stop bank robberies in the United States?

In 2014 there were almost 4,000 of them at known locations, that is to say banks. If the government can’t stop robberies/attacks at known locations, how do they propose to stop terrorist attacks which can occur anywhere?

Just one fact, US bank robberies and a little creative thinking, would enable your journalists to pierce the foggy posturing on Paris and any future or past terror attacks.

The true answer is that you can’t. Not without monitoring everyone 24/7 as far as location, conversations, purchases, etc. But so far, no reporter has forced that admission from anyone. Curious don’t you think?

Bank Crime Statistics 2014.

November 12, 2015

Quartz to open source two mapping tools

Filed under: Journalism,Mapping,Maps,Open Source,Reporting — Patrick Durusau @ 3:53 pm

Quartz to open source two mapping tools by Caroline Scott.

From the post:

News outlet Quartz is developing a searchable database of compiled map data from all over the world, and a tool to help journalists visualise this data.

The database, called Mapquery, received $35,000 (£22,900) from the Knight Foundation Prototype Fund on 3 November.

Keith Collins, project lead, said Mapquery will aim to make the research stage in the creation of maps easier and more accessible, by creating a system for finding, merging and refining geographic data.

Mapquery will not be able to produce visual maps itself, as it simply provides a database of information from which maps can be created – so Quartz will also open source Mapbuilder as the “front end” that will enable journalists to visualise the data.

Quartz aims to have a prototype of Mapquery by April, and will continue to develop Mapbuilder afterwards.

That’s news to look forward to in 2016!

I’m real curious where Quartz is going to draw the boundary around “map data?” The post mentions Mapquery including “historical boundary data,” which would be very useful for some stories, but is traditional “map data.”

What if Mapquery could integrate people who have posted images with geographic locations? So a reporter could quickly access a list of potential witnesses for events the Western media doesn’t cover?

Live feeds of the results of US bombing raids against ISIS for example. (Doesn’t cover out of deference to the US military propaganda machine or for other reasons I can’t say.)

Looking forward to more news on Mapquery and Mapbuilder!

I first saw this in a tweet by Journalism Tools.

November 6, 2015

JournalismCourses.org/

Filed under: Journalism,News,Reporting — Patrick Durusau @ 11:26 am

JournalismCourses.org/

From the webpage:

Welcome to JournalismCourses.org, an online training platform of the Knight Center for Journalism in the Americas at the University of Texas at Austin.

Since 2003, our online courses have trained more than 50,000 people from 160 countries. Initially, the program was focused on online classes for small groups of journalists, mainly from Latin America and the Caribbean, but eventually the Knight Center began offering Massive Open Online Courses. It became the first program of MOOCs in the world specializing in journalism training, but it still offers courses to small groups as well. The MOOCs are free, but participants are asked to pay a small fee for a certificate of completion. Other courses are paid, but we keep the fees as low as possible in an effort to make the courses available to as many people as possible.

Our courses cover a variety of topics including investigative reporting, ethics, digital journalism techniques, election reporting, coverage of armed conflicts, computer-assisted reporting, and many others. Our MOOCs and courses for smaller groups last from four to six weeks. They are conducted completely online and taught by some of the most respected, experienced journalists and journalism trainers in the world. The courses take full advantage of multimedia. They feature video lectures, discussion forums, audio slideshows, self-paced quizzes, and other collaborative learning technologies. Our expert instructors provide a quality learning experience for journalists seeking to improve their skills, and citizens looking to become more engaged in journalism and democracy.

The courses offered on the JournalismCourses.org platform are asynchronous, so participants can log in on the days and times that are most convenient for them. Each course, however, is open just for a specific period of time and access to it is restrict to registered students.

The Knight Center has offered online courses in English, Spanish and Portuguese. Please check this site often, as we will soon announce more online courses.

For more information about the Knight Center’s Distance Learning program, please click here .

For more information about the Knight Center’s MOOCs and how they work, see our Frequently Asked Questions (FAQ) .

JournalismCourses.org/ sponsors the Introduction to Infographics and Data Visualization MOOC I just posted about but I thought it needed more than a passing mention in a post.

More courses are on the way!

Speaking of more courses, do yourself a favor and visit: Knight Center’s Distance Learning program. You won’t be disappointed, I promise.

Introduction to Infographics and Data Visualization

Filed under: Infographics,Journalism,News,Reporting,Visualization — Patrick Durusau @ 11:08 am

Introduction to Infographics and Data Visualization by Alberto Cairo.

MOOC: Time: November 16 – December 13, 2015

From the webpage:

This is Massive Open Online Course (MOOC) is offered by the Journalism and Media Studies Centre (JMSC) at the University of Hong Kong and the Knight Center at the University of Texas at Austin. This MOOC is hosted on JournalismCourses.org, the Knight Center’s distance-learning platform. It is designed primarily for journalists and the public in Asia, but is open to people from other parts of the world as well. The Knight Center’s MOOCs are free. Other online courses, with a limited number of students, have a small fee.

Goal

This course is an introduction to the basics of the visual representation of data. In this class you will learn how to design successful charts and maps, and how to arrange them to compose cohesive storytelling pieces. We will also discuss ethical issues when designing graphics, and how the principles of Graphic Design and of Interaction Design apply to the visualization of information.

The course will have a theoretical component, as we will cover the main rules of the discipline, and also a practical one: to design basic infographics and mock ups for interactive visualizations.

One hopes that given a primarily Asian audience, that successful infographics from Asian markets will be prominent in the study materials.

Thinking that discussion among the students may identify why some infographics succeed while other fail in that context. Reasoning that all cultures have preferences or dispositions that aren’t readily apparent to outsiders.

November 4, 2015

Interhacktives

Filed under: Data Mining,Journalism,News,Reporting — Patrick Durusau @ 11:18 am

Interhacktives

I “discovered” Interhactives while following a tweet on “Excel tips for journalists.” I thought it would be a short article saying “don’t” but was mistaken. 😉

Turned out to be basic advice on “using” Excel.

Moving around a bit I found an archive of “how-to” posts and other resources for digital journalists and anyone interested in finding/analyzing content on the Web.

You won’t find discussions of Lamda Architecture here but you will find nuts-an-bolts type information, ready to be put into practice.

Visit Interhacktives and pass it along to others.

I first saw this in a tweet by Journalism Tools.

October 31, 2015

Fact Checking

Filed under: Humor,News,Reporting — Patrick Durusau @ 1:50 pm

fact-checking

Perhaps humor will help keep this in mind. 😉

I saw this in a Facebook post by Paul Prescott.

October 29, 2015

NarcoData [Why Not TrollData?] + Zero Trollerance

Filed under: #gamergate,Journalism,News,Online Harassment,Reporting — Patrick Durusau @ 2:54 pm

NarcoData is a new collaboration that aims to track and visualize the drug cartels of Mexico by Laura Hazard Owen.

From the post:

NarcoData, a collaboration between Mexican digital news site Animal Politico and data journalism platform Poderopedia, launched Tuesday with a mission to shine light on organized crime and drug trafficking in Mexico.

“The Mexican state has failed in giving its citizens accurate, updated, and systematic information about the fight against organized crime,” said Dulce Ramos, editor-in-chief of Animal Politico and the general coordinator for NarcoData. “NarcoData wants to fill that empty space.”

The site examines four decades of data to explain how drug trafficking reached its current size and influence in the country. The idea for the project came about last year, when Animal Politico obtained, via the Mexican transparency act, a government chart outlining all of the criminal cells operating in the country. Instead of immediately publishing an article with the data, Animal Politico delved further to fill in the information that the document was missing.

Even a couple of months later, when the document went public and some legacy media outlets wrote articles about it and made infographics from it, “we remained sure that that document had great potential, and we didn’t want to waste it,” Ramos said. Instead, Animal Politico requested and obtained more documents and corroborated the data with information from books, magazines, and interviews.

If you are unfamiliar with the status of the drug war in Mexico, consider the following:

Mexico’s drug war is getting even worse by Jeremy Bender:

At least 60,000 people are believed to have died between 2006 and 2012 as a result of the drug war as cartels, vigilante groups, and the Mexican army and police have battled each other.

The Staggering Death Toll of Mexico’s Drug War by Jason M. Breslow:

Last week, the Mexican government released new data showing that between 2007 and 2014 — a period that accounts for some of the bloodiest years of the nation’s war against the drug cartels — more than 164,000 people were victims of homicide. Nearly 20,000 died last year alone, a substantial number, but still a decrease from the 27,000 killed at the peak of fighting in 2011.

Over the same seven-year period, slightly more than 103,000 died in Afghanistan and Iraq, according to data from the and the website .

mexico_homicides

‘Journalists are being slaughtered’ – Mexico’s problem with press freedom by Nina Lakhani.


Journalists and press freedom groups have expressed growing anger at Mexican authorities’ failure to tackle escalating violence against reporters and activists who dare to speak out against political corruption and organised crime.

Espinosa was the 13th journalist working in Veracruz to be killed since Governor Javier Duarte from the ruling Institutional Revolutionary party (PRI) came to power in 2011. According to the press freedom organisation Article 19, the state is now the most dangerous place to be a journalist in Latin America.

According to the Committee to Protect Journalists, about 90% of journalist murders in Mexico since 1992 have gone unpunished.

Patrick Timmons, a human rights expert who investigated violence against journalists while working for the UK embassy in Mexico City, said the massacre was another attempt to silence the press: “These are targeted murders which are wiping out a whole generation of critical leaders.”

Against that background of violence and terror, NarcoData emerges. Mexican journalists speak out against the drug cartels and on behalf of the people of Mexico who suffer under the cartels.

I am embarrassed to admit sharing U.S. citizenship with the organizers of South by Southwest (SXSW). Under undisclosed “threats” of violence because of panels to discuss online harassment, the SXSW organizers cancelled the panels. Lisa Vaas captures those organizers perfectly in her headline: SXSW turns tail and runs, nixing panels on harassment.

I offer thanks that the SXSW organizers were not civil rights organizers in: SXSW turns tail and runs… [Rejoice SXSW Organizers Weren’t Civil Rights Organizers] Troll Police.

NarcoData sets an example of how to respond to drug cartels or Internet trolls. Shine a bright light on them. Something the SXSW organizers were too timid to even contemplate.

Fighting Internet trolls requires more than anecdotal accounts of abuse. Imagine a TrollData database that collects data from all forms of social media, including SMS messages and email forwarded to it. So that data analytics can be brought to bear on the data with a view towards identifying trolls by their real world identities.

Limited to Twitter but a start in that direction is described in: How do you stop Twitter trolls? Unleash a robot swarm to troll them back by Jamie Bartlett.

Knowing how to deal with Internet trolls is tricky, because the separating line between offensive expression and harassment very fine, and usually depends on your vantage point. But one subspecies, the misogynist troll, has been causing an awful lot of trouble lately. Online abuse seems to accompany every woman that pops her head over the parapet: Mary Beard, Caroline Criado-Perez, Zelda Williams and so on. It’s not just the big fish, either. The non-celebs women cop it too, but we don’t hear about it. Despite near universal condemnation of this behaviour, it just seems to be getting worse.

Today, a strange and mysterious advocacy group based in Berlin called the “Peng! Collective” have launched a new way of tackling the misogynistic Twitter trolls. They’re calling it “Zero Trollerance.”

Here’s what they are doing. If a Twitter user posts any one of around one hundred preselected terms or words that are misogynistic, a bot – an automated account – spots it, and records that user’s Twitter handle in a database. (These terms, in case you’re wondering, include, but are not limited to, the following gems: #feministsareugly #dontdatesjws “die stupid bitch”, “feminazi” and “stupid whore”.)

This is the clever bit. This is a lurking, listening bot. It’s patrolling Twitter silently as we speak and taking details of the misogynists. But then there is another fleet of a hundred or so bots – I’ll call them the attack bots – that, soon after the offending post has been identified, will start auto-tweeting messages @ the offender (more on what they tweet below).

“Zero Trollerance” is a great idea and I applaud it. But it doesn’t capture the true power of data mining, which could uncover trolls that use multiple accounts, trolls that are harassing other users via other social media, not to mention being able to shine light directly on trolls in public, quite possibly the thing they fear the most.

TrollData would require high levels of security, monitoring of all public social media and the ability to accept email and SMS messages forwarded to it, governance and data mining tools.

Mexican journalists are willing to face death to populate NarcoData, what do you say to facing down trolls?


In case you want to watch or forward the Zero Trollerance videos:

Zero Trollerance Step 1: Zero Denial

Zero Trollerance Step 2: Zero Internet

Zero Trollerance Step 3: Zero Anger

Zero Trollerance Step 4: Zero Fear

Zero Trollerance Step 5: Zero Hate

Zero Trollerance Step 6: Zero Troll

October 21, 2015

The Future Of News Is Not An Article

Filed under: Journalism,News,Publishing,Reporting — Patrick Durusau @ 2:22 pm

The Future Of News Is Not An Article by Alexis Lloyd.

Alexis challenges readers to reconsider their assumptions about the nature of “articles.” Beginning with the model for articles that was taken over from traditional print media. Whatever appeared in an article yesterday must be re-created today if there is a new article on the same subject. Not surprising since print media lacks the means to transclude content from a prior article into a new one.

She saves her best argument for last:


A news organization publishes hundreds of articles a day, then starts all over the next day, recreating any redundant content each time. This approach is deeply shaped by the constraints of print media and seems unnecessary and strange when looked at from a natively digital perspective. Can you imagine if, every time something new happened in Syria, Wikipedia published a new Syria page, and in order to understand the bigger picture, you had to manually sift through hundreds of pages with overlapping information? The idea seems absurd in that context and yet, it is essentially what news publishers do every day.

While I agree fully with the advantages Alexis summarizes as Enhanced tools for journalists, Summarization and synthesis, and Adaptive Content (see her post), there are technical and non-technical roadblocks to such changes.

First and foremost, people are being paid to re-create redundant content everyday and their comfort levels, to say nothing about their remuneration for repetitive reporting of the same content will loom large in the adoption of the technology Alexis imagines.

I recall a disturbing story from a major paper where reporters didn’t share leads or research because of fear that other reporters would “scoop” them. That sort of protectionism isn’t limited to journalists. Rumor has it that Oracle sale reps refused to enter potential sales leads in a company wide database.

I don’t understand why that sort of pettiness is tolerated but be aware that it is, both in government and corporate environments.

Second and almost as importantly, Alexis needs raise the question of semantic ROI for any semantic technology. Take her point about adoption of the Semantic Web:

but have not seen universal adoption because of the labor costs involved in doing so.

To adopt a single level of semantic encoding for all content, without regard to its value, either historical or current use, is a sure budget buster. Perhaps the business community was playing closer attention to the Semantic Web than many of us thought, hence its adoption failure.

Some content may need machine driven encoding, more valuable content may require human supervision and/or encoding and some content may not be worth encoding at all. Depends on your ROI model.

I should mention that the Semantic Web manages statements about statements (in its or other semantic systems) poorly. (AKA, “facts about facts.”) Although I hate to use the term “facts.” The very notion of “facts” is misleading and tricky under the best of circumstances.

However universal (universal = among people you know) knowledge of a “fact” may seem, the better argument is that it is only a “fact” from a particular point of view. Semantic Web based systems have difficulty with such concepts.

Third, and not mentioned by Alexis, is that semantic systems should capture and preserve trails created by information explorers. Reporters at the New York Times use databases everyday, but each search starts from scratch.

If re-making redundant information over and over again is absurd, repeating the same searches (more or less successfully) over and over again is insane.

Capturing search trails as data would enrich existing databases, especially if searchers could annotate their trails and data they encounter along the way. The more intensively searched a resource becomes, the richer its semantics. As it is today, all the effort of searchers is lost at the end of each search.

Alexis is right, let’s stop entombing knowledge in articles, papers, posts and books. It won’t be quick or easy, but worthwhile journeys rarely are.

I first saw this in a tweet by Tim Strehle.

October 13, 2015

Data Journalism Tools

Filed under: Journalism,News,Reporting,Researchers — Patrick Durusau @ 6:48 pm

Data Journalism Tools

From the webpage:

This Silk is a structured database listing tools and resources that (data) journalists might want to include in their toolkit. We tried to cover the main steps of the ddj process: from data collection and scraping to data cleaning and enhancement; from analysis to data visualization and publishing. We’re trying to showcase especially tools that are free/freemium and open source, but you will find a bit of everything.

This Silk is updated regularly: we have collected a list of hundreds of tools, which we manually tag (are they open source tools? Free? for interactive datavizs?). Make sure you follow this Silk, so you won’t miss an update!

As of 13 October 2015, there are 120 tools listed.

Graphics have a strong showing but not overly so. There are tools for collaboration, web scrapping, writing, etc.

Pitched toward journalists but librarians, researchers, bloggers, etc., will all find tools of interest at this site.

October 10, 2015

Journalism Books

Filed under: Journalism,News,Reporting — Patrick Durusau @ 8:24 pm

Journalism Books

A collection of books on journalism, said to:

…inform & inspire the future of journalism.

Some twenty (20) books so it isn’t overwhelming like some of the cybersecurity sites that push current and out-dated material with equal enthusiasm.

Journalists find and report information that some people would prefer they didn’t.

That’s buys them a lot of street cred in my book.

Despite having relatives who are journalists I have read only some of these books. Makes a nice reading list for long winter nights! (Well, aside from XQuery archives and that sort of thing.) 😉

October 8, 2015

Tipsheets & Links from GIJC15

Filed under: Journalism,News,Reporting — Patrick Durusau @ 11:00 am

Tipsheets & Links from GIJC15

Program listing from the recent Investigative Reporters and Editors (IRE) conference, annotated with tipsheets and links.

No response on my question on whether anyone is creating a subject index to tip sheets.

That’s unfortunate. The information trapped in some of these tip sheets merits wider dispersal and use, to say nothing of maintenance.

95 tools for investigative journalists

Filed under: Journalism,News,Reporting — Patrick Durusau @ 8:57 am

95 tools for investigative journalists from @Journalism2ls

95 Resources that cover:

  • Alerts
  • Analytics
  • Collect Data
  • Data Stories
  • Interactive Video
  • Location
  • Map Stories
  • Monitor News
  • Multimedia
  • People & Paper Trail
  • Privacy
  • Production
  • Reporting
  • Snowfalling
  • Social Media
  • Verification
  • Wikipedia

I re-ordered the categories into alphabetical order. In the original post, both the categories and their contents appear in no order that is apparent to me. (If you see an ordering principle in the post, please give a shout.)

Impressive collection of tools!

October 3, 2015

Computation + Journalism Symposium 2015

Filed under: Journalism,News,Reporting — Patrick Durusau @ 8:02 pm

Computation + Journalism Symposium 2015

From the webpage:

Data and computation drive our world, often without sufficient critical assessment or accountability. Journalism is adapting responsibly—finding and creating new kinds of stories that respond directly to our new societal condition. Join us for a two-day conference exploring the interface between journalism and computing.

Papers are up! Papers are up!

http://cj2015.brown.columbia.edu/papers.html

Many excellent papers but one caught my eye in particular:

DeScipher: A Text Simplification Tool for Science Journalism, Yea Seul Kim, Jessica Hullman and Eytan Adar.

High on my reading list after spending a day with “almost” explanations in technical documentation.

This could be very useful for anyone authoring useful technical documentation, not to mention writing for the general public.

September 29, 2015

BBC News Labs Project List

Filed under: Journalism,News,Reporting — Patrick Durusau @ 2:55 pm

Not only does the BBC News Lab have a cool logo:

bbc-lab-logo

They have an impressive list of projects as well:

BBC News Labs Project List

  • News Switcher – News Switcher lets BBC journalists easily switch between the differents editions of the News website
  • Pool of Video – BBC News Labs is looking into some new research questions based on AV curation.
  • Suggestr – connecting the News industry with BBC tags – This prototype, by Outlandish.com for BBC News Labs, is about enabling local News organisations to tag with BBC News tags
  • Linked data on the TV – LDPTV is a project for surfacing more News content via smart TVs
  • #newsHACK – Industry Collaboration – #newsHACK is all about intense multi-discipline collaboration on Journalism innovation.
  • BBC Rewind Collaboration – The News Labs team is working with BBC Rewind – a project liberating the BBC archive – to share tech and approaches.
  • The News Slicer – The News Slicer takes MOS running orders and transcripts from BBC Broadcast Playout, then auto tags and auto segments the stories.
  • News Rig – The future of multilingual workflow – A prototype workflow for reversioning content into multiple languages, and controlling an "on demand" multilingual news service.
  • Atomised News – with BBC R&D – News Labs has been working with BBC R&D to explore a mobile-focussed, structured breadth & depth approach to News experiences
  • Connected Studio – World Service – A programme of international innovation activities, aimed at harnessing localised industry talent to explore new opportunites.
  • Language Technology – BBC News Labs started a stream of Language Technology projects in 2014, in order to scale our storytelling globally
  • Blue Labs – BBC Blue Room and News Labs are working together to more efficiently demonstrate innovation opportunities with emerging consumer Tech.
  • Immersive News – 360 Video & VR – We are looking into the craft and applications around 360 video filming and VR for Immersive News.
  • The Journalist Toolbox – In June 2015, a hack team at #newsHACK created a concept called The Journalist Toolbox, proposing that newsroom content publishing & presentation tools needed to be more accessible for journalists.
  • SUMMA – Scalable Understanding of Multilingual MediA – This Big Data project will leverage machines to do the heavy lifting in multilingual media monitoring.
  • Window on the Newsroom – Window on the Newsroom is a prototype discovery interface to help Journalists look quickly across Newsroom system by story or metadata properties
  • The Juicer – The Juicer takes news content, automatically semantically tags it, then provides a fully featured API to access this content and data.

This list is current as of 29 September 2015 so check with the project page for new and updated project information from the BBC.

The BBC is a British institution that merits admiration. Counting English, its content is available in twenty-eight (28) languages.

Alas, “breathless superlative American English,” a staple of the Fox network, is not one of them.

« Newer PostsOlder Posts »

Powered by WordPress