Another Word For It Patrick Durusau on Topic Maps and Semantic Diversity

November 28, 2015

The First Draft Toolbox for newsgathering and verification

Filed under: Journalism,News,Reporting — Patrick Durusau @ 9:22 pm

If you are not Donald Trump or some other form of a pathological liar, then you will enjoy: The First Draft Toolbox for newsgathering and verification by Alastair Reid.

From the post:

Welcome to the First Draft Toolbox, a list of tools and sites recommended by the First Draft Coalition to help in social newsgathering, verification and more.

We will be updating the page regularly with new tools as well as more detailed explainers and guides of those listed here already. If you have any suggestions of something we may have missed or are launching a tool you think should be featured here, please let us know by emailing our editor Alastair Reid.

You can also get email alerts for when we update the page using ChangeDetection or other available tools.

So many options can be overwhelming though, and putting them into practice can be daunting when just starting out. The best advice has always been to experiment with everything but find the tools that work for you, and keep up with thought leaders and case studies to see what the experts use and how they use them.

By rough count I make it thirty-eight separate resources for newsgathering and verification. The big categories are: Social newsgathering and search tools, Location checking tools, Source verification, Image verification, YouTube Data Viewer and, Translation.

An impressive collection, several new to me and more than you will probably use at any one time. Try the most needed ones first and then branch out. Over time you will develop favorites and skill at using them.

The one omission that surprised me was Alastair failing to mention Snopes.com.

Snopes.com is one of the premier debunking sites on the WWW. For example:

Undercover Parcel Service No, UPS isn’t smuggling refugees into the United States in the dead of night.

Cetacean Harvestation No, cranberry farmers aren’t netting and canning dolphins during the harvest season.

Does that help explain Donald Trump’s standings in the polls?

Ask not only whether statements are “true,” but also what the speaker has to gain from giving them to you?

November 26, 2015

What Should the Media Do When Donald Trump Blatantly Lies? [Try Not Reporting Lies]

Filed under: Journalism,News,Reporting — Patrick Durusau @ 8:07 pm

What Should the Media Do When Donald Trump Blatantly Lies? by Matthew Ingram.

From the post:

Political speech is a unique animal, especially during election season. It often mixes hyperbole with flowery language and aggressive rhetoric designed to inflame a particular passion. But Republican presidential candidate Donald Trump is arguably in a category unto himself. More than almost any other 2016 candidate, he is prone to telling flat-out lies, making up facts, and distorting the truth to a prodigious extent.

This kind of behavior creates a tricky problem for the press. How should media companies deal with Trump and his falsehoods? If he were just a joke candidate without a hope of ever being the Republican nominee, it would be easy enough to ignore him. But he appears to stand a better than even chance of getting the nomination — he has been leading in the polls for months.

If media outlets attack Trump’s lying directly, they run the risk of being accused of bias by his supporters and Republicans in general. In fact, that kind of reaction is already occurring in response to a New York Times editorial that accused the billionaire businessman of playing fast and loose with the truth on a number of issues, including whether Muslims in New Jersey cheered the Sept. 11, 2001 terrorist attacks.

Part of the problem is that Trump and his candidacy are to some extent a creation of the mainstream media. At the very least, the two have developed a disturbingly co-dependent relationship.

As disturbing as the article in on media coverage of lies by Donald Trump, the crux of the dilemma was put this way:

since the U.S. news media is based on the commercial model—and more eyeballs on the page or the screen is good for business—the networks love it when someone like Donald Trump says outrageous stuff. Fact-checking rains on the parade of that revenue model.

Perhaps news rooms need a new version of First they came for:

First Trump lied about the refugees, and I reported it—
Because I was not a refugee.

Then Trump lied about blacks, and I reported it—
Because I was not black.

Then Trump lied about Jews, and I reported it—
Because I was not a Jew.

Donald Trump lied his way into the Whitehouse, and I made it possible-
Because fact checking conflicted with the bottom-line.

When I think about journalists who risk their lives reporting on drug cartels and violent governments, I wonder what they must think of the moral cowardice of political coverage in the United States?

November 21, 2015

Leaking Classified Information

Filed under: Cybersecurity,Journalism,News,Reporting — Patrick Durusau @ 8:57 pm

I saw a tweet recently extolling the number of classified documents that could have obtained.

Not obtaining and/or leaking classified documents of any government denies the public information it can use.

Two suggestions:

If you can obtain classified information, do.

If you have classified information, leak it in its entirety.

Before some ambitious assistant US attorney decides I am advocating illegal activity, recall that some leaks of classified information are in fact authorized by the executive branch of the United States government. Read All Leaks Are Illegal, but Some Leaks Are More Illegal Than Others by Conor Friedersdorf for some example cases.

Classification is used to conceal embarrassing information or failures. No government has a right to conceal embarrassing information or failures.

Agonizing over what to leak creates power for those with leaked information from a government. Do you see yourself as that petty and vain?

Just leak it. Let the chips fall where they may.

The history of leaking is on the side of no harm to anyone.

Start with the Pentagon Papers (U.S. Archives), Watergate at 40, Public Library of US Diplomacy, which also includes Cablegate, the Kissinger cables and Carter cables parts 1 and 2, Afghan War Diaries, the Snowden leaks and count the bodies.

So far, I’ve got nothing. Zero. The empty set.

Over forty years of leaking and no bodies. If there was even one, it would be front and center at every leak story.

Doesn’t that tell you something about the truthfulness of government objections to leaks?

November 20, 2015

Four free online plagiarism checkers

Filed under: Journalism,News,Plagiarism,Reporting — Patrick Durusau @ 5:32 pm

Four free online plagiarism checkers

From the post:

“Detecting duplicate content online has become so easy that spot-the-plagiarist is almost a party game,” former IJNet editor Nicole Martinelli wrote in 2012. “It’s no joke, however, for news organizations who discover they have published copycat content.”

When IJNet first ran Martinelli’s post, “Five free online plagiarism checkers,” two prominent U.S. journalists had recently been caught in the act: Fareed Zakaria and Jonah Lehrer.

Following acknowledgement that he had plagiarized sections of an article about gun control, Time and CNN suspended Zakaria. Lehrer first came under scrutiny for “self-plagiarism” at The New Yorker. Later, a journalist revealed Lehrer also fabricated or changed quotes attributed to Bob Dylan in his book, “Imagine.”

To date, Martinelli’s list of free plagiarism checkers has been one of IJNet’s most popular articles across all languages. It’s clear readers want to avoid the pitfalls of plagiarism, so we’ve updated the post with four of the best free online plagiarism checkers available to anyone, revised for 2015:

Great resource for checking your content and that of others for plagiarism.

The one caveat I offer is to not limit the use of text similarity software solely to plagiarism.

Text similarity can be a test for finding content that you would not otherwise discover. Depends on how high you set the test for “similarity.”

And/or it may find content that is so similar, while not plagiarism (say multiple outlets writing from the same wire service) it isn’t worth the effort to read every story that repeats the same story with some minor edits.

Multiple stories but only one wire service source. In that sense, a “plagiarism” checker can enable you to skip duplicative content.

The post I quote above was published by the international journalist’s network (ijnet). Even if you aren’t a journalist, great source to follow for developing news technology.

November 19, 2015

In the realm of verification, context is king

Filed under: Journalism,News,Reporting — Patrick Durusau @ 4:45 pm

In the realm of verification, context is king by Fergus Bell.

From the post:

By thinking about the wider context around shared UGC you can often avoid a lengthy forensic verification process where it isn’t required. For publishers looking at how they tackle competition with platforms – it is easy. Context is where you can make a distinction through strong editorial work and storytelling.

Fergus has four quick tips that will help you fashion a context for user-generated content (UGC).

Content always has a context. If you don’t supply one, consumers will invent a context for your content. (They may anyway but you can at least take the first shot at it.)

It is interesting that user-generated content (UGC) isn’t held in high regard, yet news outlets parrot the latest rantings of elected officials and public figures as gospel.

When public statements are false, such as suggesting that Syrian refugees pose a danger of terrorism, why aren’t those statements simply ignored? Why mis-inform the public?

November 18, 2015

Antidote to Network News Reporting

Filed under: Journalism,News — Patrick Durusau @ 5:35 pm

Public Trust Through Public Access to CRS Reports by Rep. Mike Quigley.

Rep. Quigley addresses Congress to urge support for House Resolution 34 saying in part:

When the average American wants to learn about a policy, where do they turn for information?

Often, the answer is the 24-hour news cycle. Often filled by talking heads and sensationalism,

Or social media and message boards, where anyone can post anything – credible or completely misinformed.

The American public is no longer being informed by the likes of Cronkite and Murrow, and it is making our public debate increasingly partisan, polarized and misinformed.

What few realize, or like to admit, is that there is a way Congress can help elevate the debate and educate our constituents with neutral, unbiased, non-partisan information from the Congressional Research Service, or CRS.

For over 100 years, CRS has served Congress’ publicly-funded think tank.

Because they serve policy-makers on both sides of the aisle, CRS researchers produce exemplary work that is accurate, non-partisan, and easy to understand.

Despite the fact that CRS receives over $100 million from taxpayers each year, its reports are not made available to the public.

Instead, constituents must request individual reports through a Congressional office.

Rep. Quigley goes on to make several public policy point in favor of House Resolution 34 but he had me at:

  1. Citizens pay for it.
  2. Citizens can’t access it online.

Citizens of the United States are paying for some the best research in the world but can’t access it online.

That is wrong on so many levels that I don’t think it needs much discussion or debate.

All U.S. citizens need to contact their representative to urge support for House Resolution 34.

Today!

PS: Congressional Research Service (CRS) reports don’t look like coiffed news anchors but then you won’t find rank speculation, rumor and falsehoods reported as facts. It’s a trade-off I’m willing to make.

November 17, 2015

Lies, Damn Lies, and Viral Content [I Know a Windmill When I See One]

Filed under: Journalism,News,Reporting — Patrick Durusau @ 7:53 pm

Lies, Damn Lies, and Viral Content How News Websites Spread (and Debunk) Online Rumors, Unverified Claims and Misinformation by Craig Silverman.

From the executive summary:

News organizations are meant to play a critical role in the dissemination of quality, accurate information in society. This has become more challenging with the onslaught of hoaxes, misinformation, and other forms of inaccurate content that flow constantly over digital platforms.

Journalists today have an imperative—and an opportunity—to sift through the mass of content being created and shared in order to separate true from false, and to help the truth to spread.

Unfortunately, as this paper details, that isn’t the current reality of how news organizations cover unverified claims, online rumors, and viral content. Lies spread much farther than the truth, and news organizations play a powerful role in making this happen.

News websites dedicate far more time and resources to propagating questionable and often false claims than they do working to verify and/or debunk viral content and online rumors. Rather than acting as a source of accurate information, online media frequently promote misinformation in an attempt to drive traffic and social engagement.

The above conclusions are the result of several months spent gathering and analyzing quantitative and qualitative data about how news organizations cover unverified claims and work to debunk false online information. This included interviews with journalists and other practitioners, a review of relevant scientific literature, and the analysis of over 1,500 news articles about more than 100 online rumors that circulated in the online press between August and December of 2014.

Many of the trends and findings detailed in the paper reflect poorly on how online media behave. Journalists have always sought out emerging (and often unverified) news. They have always followed-on the reports of other news organizations. But today the bar for what is worth giving attention seems to be much lower. There are also widely used practices in online news that are misleading and confusing to the public. These practices reflect short-term thinking that ultimately fails to deliver the full value of a piece of emerging news.

Silverman writes a compelling account (at length, some 164 pages including endnotes) to prove:

News websites dedicate far more time and resources to propagating questionable and often false claims than they do working to verify and/or debunk viral content and online rumors. Rather than acting as a source of accurate information, online media frequently promote misinformation in an attempt to drive traffic and social engagement.

We have all had the experience of watching news reports where we know the “facts” and see reporters making absurd claims about our domain of expertise. But their words may be reaching millions and you can only complain to your significant other.

I fully understand Silverman’s desire to make news reporting better, just as I labor to impress upon standards editors the difference between a reference (is used in the standard itself) and further reading (as listed in a bibliography). That distinction seems particularly difficult for some reason.

The reason I mention windmills in my title is because Silverman offers this rationale for improving verification by news outlets:

Another point of progress for journalists includes prioritizing verification and some kind of value-add to rumors and claims before engaging in propagation. This, in many cases, requires an investment of minutes rather than hours, and it helps push a story forward. The practice will lead to debunking false claims before they take hold in the collective consciousness. It will lead to fewer misinformed readers. It will surface new and important information faster. Most importantly, it will be journalism.

The benefits are:

  1. Debunking false claims before they take hold in the collective consciousness
  2. Fewer misinformed readers
  3. Surface new and important information faster
  4. It will be journalism

Starting from the top: Debunking false claims before they take hold in the collective consciousness.

How does “debunking false claims” impact traffic and social engagement? If my news outlet doesn’t have the attention grabbing headline about an image of Mary in a cheese sandwich, don’t I lose that traffic? Do you seriously think that debunking stories have the audience share of fantastic claim stories?

I suppose if the debunking involved “proving” that the image of Mary was due to witchcraft, that might drive traffic but straight up debunking seems unlikely to do so.

The second benefit was Fewer misinformed readers.

I’m at a loss to say how “fewer misinformed readers” is going to benefit the news outlet? The consequences of being misinformed accrue to the reader and not to the news outlet. I suspect the average attention span is short enough that news outlets could take the other side tomorrow without readers being overly disturbed. They would just be misinformed in a different direction.

The benefit of Surface new and important information faster comes in third.

I can see that argument but that presumes that news outlets want to report “new and important information” in the first place. What Silverman successfully argues is the practice is to report news that drives traffic and social engagement. Being “new and important has only a tangential relationship to traffic and engagement.

You probably remember during the wall-to-wall reporting about Katrina or the earthquakes in Haiti the members of the news media interviewing each other. That was nearly negative content. Even rumors and lies would be have been better.

The final advantage Silverman cites is It will be journalism.

As I said, I’m not unsympathetic to Silverman but when was journalism ever concerned with not reporting questionable and false claims? During the American Revolution perhaps?, The Civil War?, WWI?, WWWII?, Korea?, Vietnam?, and the list goes on.

There have been “good” journalists (depending upon your point of view) and “bad” journalists (again depending on your point of view). Yet, journalism, just like theology, has survived being populated in part by scalawags, charlatans, and rogues.

What journalism needs is pro-active readers to rebel against superficial, inaccurate and misleading reporting. Voting with their feet will be far more effective than exhortations to do better.

November 16, 2015

Connecting News Stories and Topic Maps

Filed under: Journalism,Marketing,News,Topic Maps — Patrick Durusau @ 2:39 pm

New WordPress plug-in Catamount aims to connect data sets and stories by Mădălina Ciobanu.

From the post:

Non-profit news organisation VT Digger, based in the United States, is building an open-source WordPress plug-in that can automatically link news stories to relevant information collected in data sets.

The tool, called Catamount, is being developed with a $35,000 (£22,900) grant from Knight Foundation Prototype Fund, and aims to give news organisations a better way of linking existing data to their daily news coverage.

Rather than hyperlinking a person’s name in a story and sending readers to a different website, publishers can use the open-source plug-in to build a small window that pops up when readers hover over a selected section of the text.

“We have this great data set, but if people don’t know it exists, they’re not going to be racing to it every single day.

“The news cycle, however, provides a hook into data,” Diane Zeigler, publisher at VT Digger, told Journalism.co.uk.

If a person is mentioned in a news story and they are also a donor, candidate or representative of an organisation involved in campaign finance, for example, an editor would be able to check the two names coincide, and give Catamount permission to link the individual to all relevant information that exists in the database.

A brief overview of this information will then be available in a pop-up box, which readers can click in order to access the full data in a separate browser window or tab.

“It’s about being able to take large data sets and make them relevant to a daily news story, so thinking about ‘why does it matter that this data has been collected for years and years’?

“In theory, it might just sit there if people don’t have a reason to draw a connection,” said Zeigler.

While Catamount only works with WordPress, the code will be made available for publishers to customise and integrate with their own content management systems.

VTDigger.org reports on the grant and other winners in Knight Foundation awards $35,000 grant to VTDigger.

Assuming that the plugin will be agnostic as to the data source, this looks like an excellent opportunity to bind topic map managed content to news stories.

You could, I suppose, return one of those dreary listings of all the prior related stories from a news source.

But that is always a lot of repetitive text to wade through for very little gain.

If you curated content with a topic map, excerpting paragraphs from prior stories when necessary for quotes, that would be a high value return for a user following your link.

Since the award was made only days ago I assume there isn’t much to be reported on the Catamount tool, as of yet. I will be following the project and will report back when something testable surfaces.

I first saw this story in an alert from Journalism.co.uk. If you aren’t already following them you should be.

November 15, 2015

On-Demand Data Journalism Training Site [Free Access Ends Nov. 30th]

Filed under: Journalism,News,Reporting — Patrick Durusau @ 10:18 pm

Investigative Reporters and Editors launches on-demand data journalism training site

From the post:

Want to become a data journalist? You’re going to need a lot of perseverance — as well as the right training.

To help make data journalism more accessible to all, Investigative Reporters and Editors (IRE) recently launched NICAR-Learn, an online platform of training videos that can be accessed from anywhere, at any time.

“NICAR-Learn is a place for journalists to demonstrate their best tricks and strategies for working with data and for others to learn from some of the best data journalists in the business,” IRE wrote in a statement.

Unlike many online training platforms, NICAR-Learn’s content won’t consist of hour-long webinars. Instead, NICAR-Learn will produce a library of short videos, often less than 10 minutes long, to train journalists on specific topics or techniques relating to data journalism.

The first NICAR-Learn videos come from data journalist MaryJo Webster, who has produced four tutorials that draw from her popular “Excel Magic” course. Users can request specific tutorials by submitting their ideas to IRE.

These videos will be available at no charge to non-IRE members through the end of November. Beginning in December, IRE will add more videos to NICAR-Learn and place them behind a paywall.

To learn more, visit NICAR-Learn’s “Getting Started” page.

I can’t say that I like “paywalls,” which I prefer to call “privilegewalls.”

Privilegewalls because that is exactly what paywalls are meant to be. To create a feeling of privilege among those who have access, to separate them from those who don’t.

And beyond a feeling of privilege, privilegewalls are meant to advantage insiders over those unfortunate enough to be outsiders. Whether those advantages are real or in the imagination of members I leave for you to debate.

Personally I think helping anyone interested to become a better journalist or data journalist will benefit everyone. Journalists, member of the public who read their publications, perhaps even the profession itself.

Here’s an example of where being a better “data journalist” would make a significant difference:

So far as I know no journalist, despite several Republican and Democratic presidential candidate debates has every asked how they propose stop bank robberies in the United States?

In 2014 there were almost 4,000 of them at known locations, that is to say banks. If the government can’t stop robberies/attacks at known locations, how do they propose to stop terrorist attacks which can occur anywhere?

Just one fact, US bank robberies and a little creative thinking, would enable your journalists to pierce the foggy posturing on Paris and any future or past terror attacks.

The true answer is that you can’t. Not without monitoring everyone 24/7 as far as location, conversations, purchases, etc. But so far, no reporter has forced that admission from anyone. Curious don’t you think?

Bank Crime Statistics 2014.

November 12, 2015

Quartz to open source two mapping tools

Filed under: Journalism,Mapping,Maps,Open Source,Reporting — Patrick Durusau @ 3:53 pm

Quartz to open source two mapping tools by Caroline Scott.

From the post:

News outlet Quartz is developing a searchable database of compiled map data from all over the world, and a tool to help journalists visualise this data.

The database, called Mapquery, received $35,000 (£22,900) from the Knight Foundation Prototype Fund on 3 November.

Keith Collins, project lead, said Mapquery will aim to make the research stage in the creation of maps easier and more accessible, by creating a system for finding, merging and refining geographic data.

Mapquery will not be able to produce visual maps itself, as it simply provides a database of information from which maps can be created – so Quartz will also open source Mapbuilder as the “front end” that will enable journalists to visualise the data.

Quartz aims to have a prototype of Mapquery by April, and will continue to develop Mapbuilder afterwards.

That’s news to look forward to in 2016!

I’m real curious where Quartz is going to draw the boundary around “map data?” The post mentions Mapquery including “historical boundary data,” which would be very useful for some stories, but is traditional “map data.”

What if Mapquery could integrate people who have posted images with geographic locations? So a reporter could quickly access a list of potential witnesses for events the Western media doesn’t cover?

Live feeds of the results of US bombing raids against ISIS for example. (Doesn’t cover out of deference to the US military propaganda machine or for other reasons I can’t say.)

Looking forward to more news on Mapquery and Mapbuilder!

I first saw this in a tweet by Journalism Tools.

November 6, 2015

JournalismCourses.org/

Filed under: Journalism,News,Reporting — Patrick Durusau @ 11:26 am

JournalismCourses.org/

From the webpage:

Welcome to JournalismCourses.org, an online training platform of the Knight Center for Journalism in the Americas at the University of Texas at Austin.

Since 2003, our online courses have trained more than 50,000 people from 160 countries. Initially, the program was focused on online classes for small groups of journalists, mainly from Latin America and the Caribbean, but eventually the Knight Center began offering Massive Open Online Courses. It became the first program of MOOCs in the world specializing in journalism training, but it still offers courses to small groups as well. The MOOCs are free, but participants are asked to pay a small fee for a certificate of completion. Other courses are paid, but we keep the fees as low as possible in an effort to make the courses available to as many people as possible.

Our courses cover a variety of topics including investigative reporting, ethics, digital journalism techniques, election reporting, coverage of armed conflicts, computer-assisted reporting, and many others. Our MOOCs and courses for smaller groups last from four to six weeks. They are conducted completely online and taught by some of the most respected, experienced journalists and journalism trainers in the world. The courses take full advantage of multimedia. They feature video lectures, discussion forums, audio slideshows, self-paced quizzes, and other collaborative learning technologies. Our expert instructors provide a quality learning experience for journalists seeking to improve their skills, and citizens looking to become more engaged in journalism and democracy.

The courses offered on the JournalismCourses.org platform are asynchronous, so participants can log in on the days and times that are most convenient for them. Each course, however, is open just for a specific period of time and access to it is restrict to registered students.

The Knight Center has offered online courses in English, Spanish and Portuguese. Please check this site often, as we will soon announce more online courses.

For more information about the Knight Center’s Distance Learning program, please click here .

For more information about the Knight Center’s MOOCs and how they work, see our Frequently Asked Questions (FAQ) .

JournalismCourses.org/ sponsors the Introduction to Infographics and Data Visualization MOOC I just posted about but I thought it needed more than a passing mention in a post.

More courses are on the way!

Speaking of more courses, do yourself a favor and visit: Knight Center’s Distance Learning program. You won’t be disappointed, I promise.

Introduction to Infographics and Data Visualization

Filed under: Infographics,Journalism,News,Reporting,Visualization — Patrick Durusau @ 11:08 am

Introduction to Infographics and Data Visualization by Alberto Cairo.

MOOC: Time: November 16 – December 13, 2015

From the webpage:

This is Massive Open Online Course (MOOC) is offered by the Journalism and Media Studies Centre (JMSC) at the University of Hong Kong and the Knight Center at the University of Texas at Austin. This MOOC is hosted on JournalismCourses.org, the Knight Center’s distance-learning platform. It is designed primarily for journalists and the public in Asia, but is open to people from other parts of the world as well. The Knight Center’s MOOCs are free. Other online courses, with a limited number of students, have a small fee.

Goal

This course is an introduction to the basics of the visual representation of data. In this class you will learn how to design successful charts and maps, and how to arrange them to compose cohesive storytelling pieces. We will also discuss ethical issues when designing graphics, and how the principles of Graphic Design and of Interaction Design apply to the visualization of information.

The course will have a theoretical component, as we will cover the main rules of the discipline, and also a practical one: to design basic infographics and mock ups for interactive visualizations.

One hopes that given a primarily Asian audience, that successful infographics from Asian markets will be prominent in the study materials.

Thinking that discussion among the students may identify why some infographics succeed while other fail in that context. Reasoning that all cultures have preferences or dispositions that aren’t readily apparent to outsiders.

November 4, 2015

Interhacktives

Filed under: Data Mining,Journalism,News,Reporting — Patrick Durusau @ 11:18 am

Interhacktives

I “discovered” Interhactives while following a tweet on “Excel tips for journalists.” I thought it would be a short article saying “don’t” but was mistaken. 😉

Turned out to be basic advice on “using” Excel.

Moving around a bit I found an archive of “how-to” posts and other resources for digital journalists and anyone interested in finding/analyzing content on the Web.

You won’t find discussions of Lamda Architecture here but you will find nuts-an-bolts type information, ready to be put into practice.

Visit Interhacktives and pass it along to others.

I first saw this in a tweet by Journalism Tools.

October 29, 2015

Parsing Academic Articles on Deadline

Filed under: Journalism,Natural Language Processing,News,Parsing — Patrick Durusau @ 8:10 pm

A group of researchers is trying to help science journalists parse academic articles on deadline by Joseph Lichterman.

From the post:

About 1.8 million new scientific papers are published each year, and most are of little consequence to the general public — or even read, really; one study estimates that up to half of all academic studies are only read by their authors, editors, and peer reviewers.

But the papers that are read can change our understanding of the universe — traces of water on Mars! — or impact our lives here on earth — sea levels rising! — and when journalists get called upon to cover these stories, they’re often thrown into complex topics without much background or understanding of the research that led to the breakthrough.

As a result, a group of researchers at Columbia and Stanford are in the process of developing Science Surveyor, a tool that algorithmically helps journalists get important context when reporting on scientific papers.

“The idea occurred to me that you could characterize the wealth of scientific literature around the topic of a new paper, and if you could do that in a way that showed the patterns in funding, or the temporal patterns of publishing in that field, or whether this new finding fit with the overall consensus with the field — or even if you could just generate images that show images very rapidly what the huge wealth, in millions of articles, in that field have shown — [journalists] could very quickly ask much better questions on deadline, and would be freed to see things in a larger context,” Columbia journalism professor Marguerite Holloway, who is leading the Science Surveyor effort, told me.

Science Surveyor is still being developed, but broadly the idea is that the tool takes the text of an academic paper and searches academic databases for other studies using similar terms. The algorithm will surface relevant articles and show how scientific thinking has changed through its use of language.

For example, look at the evolving research around neurogenesis, or the growth of new brain cells. Neurogenesis occurs primarily while babies are still in the womb, but it continues through adulthood in certain sections of the brain.

Up until a few decades ago, researchers generally thought that neurogenesis didn’t occur in humans — you had a set number of brain cells, and that’s it. But since then, research has shown that neurogenesis does in fact occur in humans.

“This tells you — aha! — this discovery is not an entirely new discovery,” Columbia professor Dennis Tenen, one of the researchers behind Science Surveyor, told me. “There was a period of activity in the ’70s, and now there is a second period of activity today. We hope to produce this interactive visualization, where given a paper on neurogenesis, you can kind of see other related papers on neurogenesis to give you the context for the story you’re telling.”

Given the number of papers published every year, an algorithmic approach like Science Surveyor is an absolute necessity.

But imagine how much richer the results would be if one of the three or four people who actually read the paper could easily link it to other research and context?

Or perhaps being a researcher who discovers the article and then blazes a trail to non-obvious literature that is also relevant?

Search engines now capture what choices users make in the links they follow but that’s a fairly crude approximation of relevancy of a particular resource. Such as not specifying why a particular resource is relevant.

Usage of literature should decide which articles merit greater attention from machine or human annotators. The last amount of humanities literature is never cited by anyone. Why spend resources annotating content that no one is likely to read?

NarcoData [Why Not TrollData?] + Zero Trollerance

Filed under: #gamergate,Journalism,News,Online Harassment,Reporting — Patrick Durusau @ 2:54 pm

NarcoData is a new collaboration that aims to track and visualize the drug cartels of Mexico by Laura Hazard Owen.

From the post:

NarcoData, a collaboration between Mexican digital news site Animal Politico and data journalism platform Poderopedia, launched Tuesday with a mission to shine light on organized crime and drug trafficking in Mexico.

“The Mexican state has failed in giving its citizens accurate, updated, and systematic information about the fight against organized crime,” said Dulce Ramos, editor-in-chief of Animal Politico and the general coordinator for NarcoData. “NarcoData wants to fill that empty space.”

The site examines four decades of data to explain how drug trafficking reached its current size and influence in the country. The idea for the project came about last year, when Animal Politico obtained, via the Mexican transparency act, a government chart outlining all of the criminal cells operating in the country. Instead of immediately publishing an article with the data, Animal Politico delved further to fill in the information that the document was missing.

Even a couple of months later, when the document went public and some legacy media outlets wrote articles about it and made infographics from it, “we remained sure that that document had great potential, and we didn’t want to waste it,” Ramos said. Instead, Animal Politico requested and obtained more documents and corroborated the data with information from books, magazines, and interviews.

If you are unfamiliar with the status of the drug war in Mexico, consider the following:

Mexico’s drug war is getting even worse by Jeremy Bender:

At least 60,000 people are believed to have died between 2006 and 2012 as a result of the drug war as cartels, vigilante groups, and the Mexican army and police have battled each other.

The Staggering Death Toll of Mexico’s Drug War by Jason M. Breslow:

Last week, the Mexican government released new data showing that between 2007 and 2014 — a period that accounts for some of the bloodiest years of the nation’s war against the drug cartels — more than 164,000 people were victims of homicide. Nearly 20,000 died last year alone, a substantial number, but still a decrease from the 27,000 killed at the peak of fighting in 2011.

Over the same seven-year period, slightly more than 103,000 died in Afghanistan and Iraq, according to data from the and the website .

mexico_homicides

‘Journalists are being slaughtered’ – Mexico’s problem with press freedom by Nina Lakhani.


Journalists and press freedom groups have expressed growing anger at Mexican authorities’ failure to tackle escalating violence against reporters and activists who dare to speak out against political corruption and organised crime.

Espinosa was the 13th journalist working in Veracruz to be killed since Governor Javier Duarte from the ruling Institutional Revolutionary party (PRI) came to power in 2011. According to the press freedom organisation Article 19, the state is now the most dangerous place to be a journalist in Latin America.

According to the Committee to Protect Journalists, about 90% of journalist murders in Mexico since 1992 have gone unpunished.

Patrick Timmons, a human rights expert who investigated violence against journalists while working for the UK embassy in Mexico City, said the massacre was another attempt to silence the press: “These are targeted murders which are wiping out a whole generation of critical leaders.”

Against that background of violence and terror, NarcoData emerges. Mexican journalists speak out against the drug cartels and on behalf of the people of Mexico who suffer under the cartels.

I am embarrassed to admit sharing U.S. citizenship with the organizers of South by Southwest (SXSW). Under undisclosed “threats” of violence because of panels to discuss online harassment, the SXSW organizers cancelled the panels. Lisa Vaas captures those organizers perfectly in her headline: SXSW turns tail and runs, nixing panels on harassment.

I offer thanks that the SXSW organizers were not civil rights organizers in: SXSW turns tail and runs… [Rejoice SXSW Organizers Weren’t Civil Rights Organizers] Troll Police.

NarcoData sets an example of how to respond to drug cartels or Internet trolls. Shine a bright light on them. Something the SXSW organizers were too timid to even contemplate.

Fighting Internet trolls requires more than anecdotal accounts of abuse. Imagine a TrollData database that collects data from all forms of social media, including SMS messages and email forwarded to it. So that data analytics can be brought to bear on the data with a view towards identifying trolls by their real world identities.

Limited to Twitter but a start in that direction is described in: How do you stop Twitter trolls? Unleash a robot swarm to troll them back by Jamie Bartlett.

Knowing how to deal with Internet trolls is tricky, because the separating line between offensive expression and harassment very fine, and usually depends on your vantage point. But one subspecies, the misogynist troll, has been causing an awful lot of trouble lately. Online abuse seems to accompany every woman that pops her head over the parapet: Mary Beard, Caroline Criado-Perez, Zelda Williams and so on. It’s not just the big fish, either. The non-celebs women cop it too, but we don’t hear about it. Despite near universal condemnation of this behaviour, it just seems to be getting worse.

Today, a strange and mysterious advocacy group based in Berlin called the “Peng! Collective” have launched a new way of tackling the misogynistic Twitter trolls. They’re calling it “Zero Trollerance.”

Here’s what they are doing. If a Twitter user posts any one of around one hundred preselected terms or words that are misogynistic, a bot – an automated account – spots it, and records that user’s Twitter handle in a database. (These terms, in case you’re wondering, include, but are not limited to, the following gems: #feministsareugly #dontdatesjws “die stupid bitch”, “feminazi” and “stupid whore”.)

This is the clever bit. This is a lurking, listening bot. It’s patrolling Twitter silently as we speak and taking details of the misogynists. But then there is another fleet of a hundred or so bots – I’ll call them the attack bots – that, soon after the offending post has been identified, will start auto-tweeting messages @ the offender (more on what they tweet below).

“Zero Trollerance” is a great idea and I applaud it. But it doesn’t capture the true power of data mining, which could uncover trolls that use multiple accounts, trolls that are harassing other users via other social media, not to mention being able to shine light directly on trolls in public, quite possibly the thing they fear the most.

TrollData would require high levels of security, monitoring of all public social media and the ability to accept email and SMS messages forwarded to it, governance and data mining tools.

Mexican journalists are willing to face death to populate NarcoData, what do you say to facing down trolls?


In case you want to watch or forward the Zero Trollerance videos:

Zero Trollerance Step 1: Zero Denial

Zero Trollerance Step 2: Zero Internet

Zero Trollerance Step 3: Zero Anger

Zero Trollerance Step 4: Zero Fear

Zero Trollerance Step 5: Zero Hate

Zero Trollerance Step 6: Zero Troll

October 21, 2015

The Future Of News Is Not An Article

Filed under: Journalism,News,Publishing,Reporting — Patrick Durusau @ 2:22 pm

The Future Of News Is Not An Article by Alexis Lloyd.

Alexis challenges readers to reconsider their assumptions about the nature of “articles.” Beginning with the model for articles that was taken over from traditional print media. Whatever appeared in an article yesterday must be re-created today if there is a new article on the same subject. Not surprising since print media lacks the means to transclude content from a prior article into a new one.

She saves her best argument for last:


A news organization publishes hundreds of articles a day, then starts all over the next day, recreating any redundant content each time. This approach is deeply shaped by the constraints of print media and seems unnecessary and strange when looked at from a natively digital perspective. Can you imagine if, every time something new happened in Syria, Wikipedia published a new Syria page, and in order to understand the bigger picture, you had to manually sift through hundreds of pages with overlapping information? The idea seems absurd in that context and yet, it is essentially what news publishers do every day.

While I agree fully with the advantages Alexis summarizes as Enhanced tools for journalists, Summarization and synthesis, and Adaptive Content (see her post), there are technical and non-technical roadblocks to such changes.

First and foremost, people are being paid to re-create redundant content everyday and their comfort levels, to say nothing about their remuneration for repetitive reporting of the same content will loom large in the adoption of the technology Alexis imagines.

I recall a disturbing story from a major paper where reporters didn’t share leads or research because of fear that other reporters would “scoop” them. That sort of protectionism isn’t limited to journalists. Rumor has it that Oracle sale reps refused to enter potential sales leads in a company wide database.

I don’t understand why that sort of pettiness is tolerated but be aware that it is, both in government and corporate environments.

Second and almost as importantly, Alexis needs raise the question of semantic ROI for any semantic technology. Take her point about adoption of the Semantic Web:

but have not seen universal adoption because of the labor costs involved in doing so.

To adopt a single level of semantic encoding for all content, without regard to its value, either historical or current use, is a sure budget buster. Perhaps the business community was playing closer attention to the Semantic Web than many of us thought, hence its adoption failure.

Some content may need machine driven encoding, more valuable content may require human supervision and/or encoding and some content may not be worth encoding at all. Depends on your ROI model.

I should mention that the Semantic Web manages statements about statements (in its or other semantic systems) poorly. (AKA, “facts about facts.”) Although I hate to use the term “facts.” The very notion of “facts” is misleading and tricky under the best of circumstances.

However universal (universal = among people you know) knowledge of a “fact” may seem, the better argument is that it is only a “fact” from a particular point of view. Semantic Web based systems have difficulty with such concepts.

Third, and not mentioned by Alexis, is that semantic systems should capture and preserve trails created by information explorers. Reporters at the New York Times use databases everyday, but each search starts from scratch.

If re-making redundant information over and over again is absurd, repeating the same searches (more or less successfully) over and over again is insane.

Capturing search trails as data would enrich existing databases, especially if searchers could annotate their trails and data they encounter along the way. The more intensively searched a resource becomes, the richer its semantics. As it is today, all the effort of searchers is lost at the end of each search.

Alexis is right, let’s stop entombing knowledge in articles, papers, posts and books. It won’t be quick or easy, but worthwhile journeys rarely are.

I first saw this in a tweet by Tim Strehle.

October 18, 2015

Tracie Powell: “We’re supposed to challenge power…

Filed under: Government,Journalism,Social Media — Patrick Durusau @ 10:01 pm

Tracie Powell: “We’re supposed to challenge power…it seems like we’ve abdicated that to social media” by Laura Hazard Owen.

From the post:

Tracie Powell tries not to use the word “diversity” anymore.

“When you talk about diversity, people’s eyes glaze over,” Powell, the founder of All Digitocracy, told me. The site covers tech, policy, and the impact of media on communities that Powell describes as “emerging audiences” — people of color and of different sexual orientations and gender identities.

I first heard Powell speak at the LION conference for hyperlocal publishers in Chicago earlier this month, where she stood in front of the almost entirely white audience to discuss how journalists and news organizations can get better at reporting for more people.

I followed up with Powell, who is currently a John S. Knight Journalism Fellow at Stanford, to hear more. “If we [as journalists] don’t do a better job at engaging with these audiences, we’re dead,” Powell said. “Our survival depends on reaching these emerging audiences.”

Here’s a lightly condensed and edited version of our conversation.

Warning: Challenging power is far more risky than supporting fiery denunciations of the most vulnerable and least powerful in society.

From women facing hard choices about pregnancy, rape victims, survivors of abuse both physical and emotional, or those who have lived with doubt, discrimination and deprivation as day to day realities, victims of power aren’t hard to find.

One of the powers that needs to be challenged is the news media itself. Take for example the near constant emphasis on gun violence and mass shootings. If you were to take the news media at face value, you would be frightened to go outside.

But, a 2013 Pew Center Report, Gun Homicide Rate Down 49% Since 1993 Peak; Public Unaware tell a different tale:

pew-gun-deaths

Not as satisfying as taking down a representative or senator but in the long run, influencing the mass media may be a more reliable path to challenging power.

October 13, 2015

Data Journalism Tools

Filed under: Journalism,News,Reporting,Researchers — Patrick Durusau @ 6:48 pm

Data Journalism Tools

From the webpage:

This Silk is a structured database listing tools and resources that (data) journalists might want to include in their toolkit. We tried to cover the main steps of the ddj process: from data collection and scraping to data cleaning and enhancement; from analysis to data visualization and publishing. We’re trying to showcase especially tools that are free/freemium and open source, but you will find a bit of everything.

This Silk is updated regularly: we have collected a list of hundreds of tools, which we manually tag (are they open source tools? Free? for interactive datavizs?). Make sure you follow this Silk, so you won’t miss an update!

As of 13 October 2015, there are 120 tools listed.

Graphics have a strong showing but not overly so. There are tools for collaboration, web scrapping, writing, etc.

Pitched toward journalists but librarians, researchers, bloggers, etc., will all find tools of interest at this site.

October 10, 2015

Journalism Books

Filed under: Journalism,News,Reporting — Patrick Durusau @ 8:24 pm

Journalism Books

A collection of books on journalism, said to:

…inform & inspire the future of journalism.

Some twenty (20) books so it isn’t overwhelming like some of the cybersecurity sites that push current and out-dated material with equal enthusiasm.

Journalists find and report information that some people would prefer they didn’t.

That’s buys them a lot of street cred in my book.

Despite having relatives who are journalists I have read only some of these books. Makes a nice reading list for long winter nights! (Well, aside from XQuery archives and that sort of thing.) 😉

October 8, 2015

Tipsheets & Links from GIJC15

Filed under: Journalism,News,Reporting — Patrick Durusau @ 11:00 am

Tipsheets & Links from GIJC15

Program listing from the recent Investigative Reporters and Editors (IRE) conference, annotated with tipsheets and links.

No response on my question on whether anyone is creating a subject index to tip sheets.

That’s unfortunate. The information trapped in some of these tip sheets merits wider dispersal and use, to say nothing of maintenance.

95 tools for investigative journalists

Filed under: Journalism,News,Reporting — Patrick Durusau @ 8:57 am

95 tools for investigative journalists from @Journalism2ls

95 Resources that cover:

  • Alerts
  • Analytics
  • Collect Data
  • Data Stories
  • Interactive Video
  • Location
  • Map Stories
  • Monitor News
  • Multimedia
  • People & Paper Trail
  • Privacy
  • Production
  • Reporting
  • Snowfalling
  • Social Media
  • Verification
  • Wikipedia

I re-ordered the categories into alphabetical order. In the original post, both the categories and their contents appear in no order that is apparent to me. (If you see an ordering principle in the post, please give a shout.)

Impressive collection of tools!

October 3, 2015

Computation + Journalism Symposium 2015

Filed under: Journalism,News,Reporting — Patrick Durusau @ 8:02 pm

Computation + Journalism Symposium 2015

From the webpage:

Data and computation drive our world, often without sufficient critical assessment or accountability. Journalism is adapting responsibly—finding and creating new kinds of stories that respond directly to our new societal condition. Join us for a two-day conference exploring the interface between journalism and computing.

Papers are up! Papers are up!

http://cj2015.brown.columbia.edu/papers.html

Many excellent papers but one caught my eye in particular:

DeScipher: A Text Simplification Tool for Science Journalism, Yea Seul Kim, Jessica Hullman and Eytan Adar.

High on my reading list after spending a day with “almost” explanations in technical documentation.

This could be very useful for anyone authoring useful technical documentation, not to mention writing for the general public.

September 29, 2015

BBC News Labs Project List

Filed under: Journalism,News,Reporting — Patrick Durusau @ 2:55 pm

Not only does the BBC News Lab have a cool logo:

bbc-lab-logo

They have an impressive list of projects as well:

BBC News Labs Project List

  • News Switcher – News Switcher lets BBC journalists easily switch between the differents editions of the News website
  • Pool of Video – BBC News Labs is looking into some new research questions based on AV curation.
  • Suggestr – connecting the News industry with BBC tags – This prototype, by Outlandish.com for BBC News Labs, is about enabling local News organisations to tag with BBC News tags
  • Linked data on the TV – LDPTV is a project for surfacing more News content via smart TVs
  • #newsHACK – Industry Collaboration – #newsHACK is all about intense multi-discipline collaboration on Journalism innovation.
  • BBC Rewind Collaboration – The News Labs team is working with BBC Rewind – a project liberating the BBC archive – to share tech and approaches.
  • The News Slicer – The News Slicer takes MOS running orders and transcripts from BBC Broadcast Playout, then auto tags and auto segments the stories.
  • News Rig – The future of multilingual workflow – A prototype workflow for reversioning content into multiple languages, and controlling an "on demand" multilingual news service.
  • Atomised News – with BBC R&D – News Labs has been working with BBC R&D to explore a mobile-focussed, structured breadth & depth approach to News experiences
  • Connected Studio – World Service – A programme of international innovation activities, aimed at harnessing localised industry talent to explore new opportunites.
  • Language Technology – BBC News Labs started a stream of Language Technology projects in 2014, in order to scale our storytelling globally
  • Blue Labs – BBC Blue Room and News Labs are working together to more efficiently demonstrate innovation opportunities with emerging consumer Tech.
  • Immersive News – 360 Video & VR – We are looking into the craft and applications around 360 video filming and VR for Immersive News.
  • The Journalist Toolbox – In June 2015, a hack team at #newsHACK created a concept called The Journalist Toolbox, proposing that newsroom content publishing & presentation tools needed to be more accessible for journalists.
  • SUMMA – Scalable Understanding of Multilingual MediA – This Big Data project will leverage machines to do the heavy lifting in multilingual media monitoring.
  • Window on the Newsroom – Window on the Newsroom is a prototype discovery interface to help Journalists look quickly across Newsroom system by story or metadata properties
  • The Juicer – The Juicer takes news content, automatically semantically tags it, then provides a fully featured API to access this content and data.

This list is current as of 29 September 2015 so check with the project page for new and updated project information from the BBC.

The BBC is a British institution that merits admiration. Counting English, its content is available in twenty-eight (28) languages.

Alas, “breathless superlative American English,” a staple of the Fox network, is not one of them.

September 11, 2015

Capturing Quotes from Video

Filed under: Editor,Journalism,News — Patrick Durusau @ 4:13 pm

A new tool from The Times of London lets you easily detect and capture quotes from a video by Shan Wang.

From the post:

“As I went through these articles and came across a text quote, I kept thinking, ‘Why can’t I just click on it and see the corresponding part of the video and get the full experience of how they said it?’”

Surfacing the latest Donald Trump gem from a long, rambling video to share it in a story can be a chore. A new tool from The Times of London called quickQuote, recently open sourced, allows users to upload a video, search for and select words and sentences from an automatically generated transcription of that video, and then embed the chosen quote with the accompanying video excerpt into any article.

Users can then highlight a quote they want to use, edit the quote on the same page to correct for any errors made by the automated transcription service, and then preview and export an embeddable quote/video clip package.

At Github: http://times.github.io/quickQuote/

This is awesome!

I know what I am going to be doing this weekend!

September 10, 2015

14 innovative journalism courses…

Filed under: Journalism,News,Reporting — Patrick Durusau @ 7:49 pm

14 innovative journalism courses to follow this Fall by Aleszu Bajak.

From the post:

With classes back in session, we wanted to highlight a few forward-looking courses being taught at journalism schools across the country. But first, to introduce these syllabi, we recommend “Those Who Do, Also Teach: David Carr’s Gift to Journalism Schools,” by Molly Wright Steenson for Storybench. It’s a look at Carr’s inspiring syllabus, Press Play, and why it resonates today more than ever. Below, an excerpt from Carr’s syllabus:

While writing, shooting, and editing are often solitary activities, great work emerges in the spaces between people. We will be working in groups with peer and teacher edits. There will be a number of smaller assignments, but the goal is that you will leave here with a single piece of work that reflects your capabilities as a maker of media. But remember, evaluations will be based not just on your efforts, but on your ability to bring excellence out of the people around you.

So take a look at the following J-school courses. Check out Robert Hernandez’s experiments in VR journalism, Molly Wright Steenson’s exploration of information architecture and the media landscape, Dan Nguyen’s data reporting class, Catherine D’Ignazio’s projects melding civic art and design, and our own Jeff Howe’s media innovation studio at Northeastern University, among many others. You don’t have to be a journalism student to dig deep into the readings, try out some assignments, and learn something new.

If you get a thrill from discovering new information or mastering a new skill, you will have a field day with this course listing.

You want to follow Storybench, which self-describes as:

Storybench is a collaboration between the Media Innovation track at Northeastern University’s School of Journalism and Esquire magazine.

At Storybench, we want to reinvigorate and reimagine what digital journalism can be. This means providing an “under the hood” look at the latest and most inventive examples of digital creativity—from data visualization projects to interactive documentaries—as well as the tools and innovators behind them.

Whether you are a veteran newsroom editor, web designer, budding coder or journalism student, Storybench will help you learn what is being built and how, so you can find your way to what might be built next.

Storybench‘s editor is Aleszu Bajak, a science journalist and former Knight Science Journalism Fellow at MIT. He is an alum of Science Friday, the founder of LatinAmericanScience.org and is passionate about breaking down the divide between journalists, developers and designers. He can be reached at aleszubajak [at] gmail or at aleszu.com.

Enjoy!

August 17, 2015

101 webscraping and research tasks for the data journalist

Filed under: Journalism,News,Python,Reporting,Web Scrapers — Patrick Durusau @ 4:56 pm

101 webscraping and research tasks for the data journalist by Dan Nguyen.

From the webpage:

This repository contains 101 Web data-collection tasks in Python 3 that I assigned to my Computational Journalism class in Spring 2015 to give them regular exercise in programming and conducting research, and to expose them to the variety of data published online.

The hard part of many of these tasks is researching and finding the actual data source. The scripts need only concern itself with fetching the data and printing the answer in the least painful way possible. Since the Computational Journalism class wasn’t intended to be an actual programming class, adherence to idioms and best codes practices was not emphasized…(especially since I’m new to Python myself!)

Too good of an idea to not steal! Practical and immediate results, introduction to coding, etc.

What 101 tasks do you want to document and with what tool?

PS: The Computational Journalism class site has a nice set of online references for Python.

August 9, 2015

User-generated content can traumatize journalists… [And the Problem Is?]

Filed under: Journalism,News,Reporting — Patrick Durusau @ 4:38 pm

User-generated content can traumatize journalists who work with it — a new project aims to help by Laura Hazard Owen.

From the post:

Journalists and human rights workers who work with troubling user-generated content as part of their jobs may experience vicarious trauma as a result of handling distressing content. A new research project aims to help by surveying and interviewing such workers and developing a set of best practices for news and humanitarian organizations.

Nonprofit think-tank Eyewitness Media Hub is running the project with backing from the Open Society Foundation. EMH was founded in 2014 by former Tow fellows Sam Dubberley, Pete Brown, and Claire Wardle, who had previously researched how broadcasters use user-generated content (UGC) in their news output, along with Jenni Sargent.

“A lot of research so far has [questioned whether] vicarious trauma is something that exists,” said Dubberley. “We’re starting from the premise that it does exist, and would like to understand what organizations are doing about it, and how people who are using it on a day-to-day basis feel about it.” As head of the Eurovision News Exchange, he said, “I had a team of 20 journalists sourcing content from Syria and the Arab Spring through YouTube and saw them being impacted by it, quite seriously.”

I am quite mystified as to why news content traumatizing journalist or members of the general pubic is a problem?

If we could vicariously experience the horrors that are financed by the United States government and others, literally be retching from fear of the next newspaper, radio broadcast or cable news broadcast, wouldn’t that be a good thing?

If anything, reporters need to take the gloves off and record death rattles, people screaming in agony, calling for death, while identifying the forces that were responsible.

One wonders how may heroes parades would happen if the streets were lined the photos of their victims, pulsing with audio of the last moments of their lives.

How proud would you feel to have butchered innocent women and children in service of your country?

Let’s bring the reality of war back to the evening dinner table. It helped end Viet-Nam. Perhaps it could help end some of the current cycle of madness.

July 21, 2015

Top Ten #ddj

Filed under: Journalism,News,Reporting — Patrick Durusau @ 4:41 pm

Top Ten #ddj: The Week’s Most Popular Data Journalism Links

From the post:

What’s the data-driven journalism crowd tweeting? Here are the top ten links for July 9 to 16: +300 sites with free geographic datasets (@sciremotesense); graphing German YouTube (@SPIEGELOLINE); Australia’s mining footprint (@ICIJorg); democratizing data (OKFN); and more.

Enjoy!

July 15, 2015

+300 Latin American Investigations

Filed under: Journalism,News,Reporting — Patrick Durusau @ 2:19 pm

Database Launched with +300 Latin American Investigations by Gabriela Manuli.

A unique database of more than 300 investigative journalism reports from across Latin America is now available from The Institute for Press and Society (Instituto Prensa y Sociedad, or IPYS). Called BIPYS (Banco de Investigaciones Periodísticas, or Bank of Investigative Journalism) the UNESCO-backed initiative was announced July 6 at the annual conference of Abraji, Brazil’s investigative journalism association.

BIPYS is a repository of many of the best examples of investigative journalism in the region, comprised largely of winners of the annual Latin American Investigative Journalism Awards that IPYS and Transparency International have given out for the past 13 years.

Investigations cover a wide range of topics, including corruption, malfeasance, organized crime, environment, national security, and human rights.

See Gabriela’s post for more but in summary the site is still under development and fees being discussed.

An admirable effort considering that words in Latin American can and do have real consequences.

Unlike some places where disagreement can be quite heated but when the broadcast ends, the participants slip away for drinks together. Meanwhile, the subjects of their disagreement continue to struggle and die due to policy decisions made far, far away.

July 13, 2015

Building News Apps Quickly?

Filed under: Journalism,News,Reporting — Patrick Durusau @ 5:55 pm

Want to make it easier to build news apps quickly? Vox Media has opensourced its solution, Autotune by Justin Ellis.

From the post:

Making a beautiful app for news is great; making a beautiful reusable app for news is better. At least that’s the thinking behind a new project released by Vox Media today: Autotune is a system meant to simplify the creation and duplication of things like data visualizations, graphics, or games.

Autotune was designed by members of the Vox Media product team to cut down on the repetitive work of taking one project — say, a an image slider — and making it easy to use elsewhere. It’s “a centralized management system for your charts, graphics, quizzes and other tools, brought to you by the Editorial Products team at Vox Media,” according to the project’s GitHub page. And, yes, that means Autotune is open source.

Sounds like a great project but I will have to get a cellphone to pass judgement on apps. 😉 I would have to get a Farraday cage to keep it in when not testing apps.

« Newer PostsOlder Posts »

Powered by WordPress