Another Word For It Patrick Durusau on Topic Maps and Semantic Diversity

December 6, 2015

Learning from Distributed Data:… [Beating the Bounds]

Filed under: Distributed Computing,Distributed Systems,Topic Maps — Patrick Durusau @ 10:35 pm

Learning from Distributed Data: Mathematical and Computational Methods to Analyze De-centralized Information.

From the post:

Scientific advances typically produce massive amounts of data, which is, of course, a good thing. But when many of these datasets are at multiple locations, instead of all in one place, it becomes difficult and costly for researchers to extract meaningful information from them.

So, the question becomes: “How do we learn from these datasets if they cannot be shared or placed in a central location?” says Trilce Estrada-Piedra.

Estrada-Piedra, an assistant professor of computer sciences at the University of New Mexico (UNM) is working to find the solution. She designs software that will enable researchers to collaborate with one another, using decentralized data, without jeopardizing privacy or raising infrastructure concerns.

“Our contributions will help speed research in a variety of sciences like health informatics, astronomy, high energy physics, climate simulations and drug design,” Estrada-Piedra says. “It will be relevant for problems where data is spread out in many different locations.”

The aim of the National Science Foundation (NSF)-funded scientist’s project is to build mathematical models from each of the “local” data banks — those at each distributed site. These models will capture data patterns, rather than specific data points.

“Researchers then can share only the models, instead of sharing the actual data,” she says, citing a medical database as an example. “The original data, for example, would have the patient’s name, age, gender and particular metrics like blood pressure, heart rate, etcetera, and that one patient would be a data point. But the models will project his or her information and extract knowledge from the data. It would just be math. The idea is to build these local models that don’t have personal information, and then share the models without compromising privacy.”

Estrada-Piedra is designing algorithms for data projections and middleware: software that acts as a bridge between an operating system or database and applications, especially on a network. This will allow distributed data to be analyzed effectively.
….

I’m looking forward to hearing more about Estrada-Piedra’s work, although we all know there are more than data projection and middleware issues involved. Those are very real and very large problems, but as with all human endeavors, the last mile is defined by local semantics.

Efficiently managing local semantics, that is enabling others to seamlessly navigate your local semantics and to in turn navigate the local semantics of others, isn’t a technical task, or at least not primarily.

The primary obstacle to such a task is captured by John D. Cook in Medieval software project management.

The post isn’t long so I will quite it here:

Centuries ago, English communities would walk the little boys around the perimeter of their parish as a way of preserving land records. This was called “beating the bounds.” The idea was that by teaching the boundaries to someone young, the knowledge would be preserved for the lifespan of that person. Of course modern geological survey techniques make beating the bounds unnecessary.

Software development hasn’t reached the sophistication of geographic survey. Many software shops use a knowledge management system remarkably similar to beating the bounds. They hire a new developer to work on a new project. That developer will remain tied to that project for the rest of his or her career, like a serf tied to the land. The knowledge essential to maintaining that project resides only in the brain of its developer. There are no useful written records or reliable maps, just like medieval property boundaries.

Does that sound familiar? That only you or another person “know” the semantics of your datastores? Are you still “beating the bounds” to document your data semantics?

Or as John puts it:

There are no useful written records or reliable maps, just like medieval property boundaries.

It doesn’t have to be that way. You could have reliable maps, reliable maps that are updated when your data is mapped for yet another project. Another ETL is the acronym.

You can, as a manager, of course, simply allow data knowledge to evaporate from your projects but that seems like a very poor business practice.

Johanna Rothman responded to John’s post in Breaking Free of Legacy Projects with the suggestion that every project should have several young boys and girls “beating the bounds” for every major project.

The equivalent of avoiding a single point of failure in medieval software project management.

Better than relying on a single programmer but using more modern information management/retention techniques would be a better option.

I guess the question is do you like using medieval project management techniques for your data or not?

If you do, you won’t be any worse off than any of your competitors with a similar policy.

On the other hand, should one of your competitors break ranks, start using topic maps for example for mission critical data, well, you have been warned.

Innovation Down Under!

Filed under: Funding,Government,NSF,Researchers — Patrick Durusau @ 9:40 pm

Twenty-nine “Welcome to the Ideas Boom” one-pagers from innovation.gov.au.

I saw this in a tweet by Leanne O’Donnell thanking @stilgherrian for putting these in one PDF file.

Hard to say what the results will be but certainly more successful than fattening the usual suspects. (NSF: BD Spokes (pronounced “hoax”) initiative)

Watch for the success factors so you can build upon the experience Australia has with its new approaches.

Racist algorithms: how Big Data makes bias seem objective

Filed under: BigData,Ethics — Patrick Durusau @ 8:46 pm

Racist algorithms: how Big Data makes bias seem objective by Cory Doctorow.

From the post:

The Ford Foundation’s Michael Brennan discusses the many studies showing how algorithms can magnify bias — like the prevalence of police background check ads shown against searches for black names.

What’s worse is the way that machine learning magnifies these problems. If an employer only hires young applicants, a machine learning algorithm will learn to screen out all older applicants without anyone having to tell it to do so.

Worst of all is that the use of algorithms to accomplish this discrimination provides a veneer of objective respectability to racism, sexism and other forms of discrimination.

Cory has a good example of “hidden” bias in data analysis and has suggestions for possible improvement.

Although I applaud the notion of “algorithmic transparency,” the issue of bias in algorithms may be more subtle than you think.

Lauren J. Young reports in Computer Scientists Find Bias in Algorithms that the bias problem can be especially acute with self-improving algorithms. Algorithms, like users have experiences and those experiences can lead to bias.

Lauren’s article is a good introduction to the concept of bias in algorithms, but for the full monty, see: Certifying and removing disparate impact by Michael Feldman, et al.

Abstract:

What does it mean for an algorithm to be biased? In U.S. law, unintentional bias is encoded via disparate impact, which occurs when a selection process has widely different outcomes for different groups, even as it appears to be neutral. This legal determination hinges on a definition of a protected class (ethnicity, gender, religious practice) and an explicit description of the process.

When the process is implemented using computers, determining disparate impact (and hence bias) is harder. It might not be possible to disclose the process. In addition, even if the process is open, it might be hard to elucidate in a legal setting how the algorithm makes its decisions. Instead of requiring access to the algorithm, we propose making inferences based on the data the algorithm uses.

We make four contributions to this problem. First, we link the legal notion of disparate impact to a measure of classification accuracy that while known, has received relatively little attention. Second, we propose a test for disparate impact based on analyzing the information leakage of the protected class from the other data attributes. Third, we describe methods by which data might be made unbiased. Finally, we present empirical evidence supporting the effectiveness of our test for disparate impact and our approach for both masking bias and preserving relevant information in the data. Interestingly, our approach resembles some actual selection practices that have recently received legal scrutiny.

Bear in mind that disparate impact is only one form of bias for a selected set of categories. And that bias can be introduced prior to formal data analysis.

Rather than say data or algorithms can be made unbiased, say rather that known biases can be reduced to acceptable levels, for some definition of acceptable.

Does Your Hello Barbie Have An STD? (IIoT)

Filed under: Cybersecurity,Privacy,Security — Patrick Durusau @ 3:36 pm

[STD = Security Transmitted Disease]

Internet-connected Hello Barbie doll gets bitten by nasty POODLE crypto bug by Dan Goodin.

From the post:

A recent review of the Internet-connected Hello Barbie doll from toymaker Mattel uncovered several red flags. Not only did the toy use a weak authentication mechanism that made it possible for attackers to monitor communications the doll sent to servers, but those servers were also vulnerable to POODLE, an attack disclosed 14 months ago that breaks HTTPS encryption.

The vulnerabilities, laid out in a report published Friday by security firm Bluebox Labs, are the latest black eye for so-called “Internet of Things” devices. The term is applied to appliances and other everyday devices that are connected to the Internet, supposedly to give them a wider range of capabilities. The Hello Barbie doll is able to hold real-time conversations by uploading the words a child says to a server. Instant processing on the server then allows the doll to provide an appropriate response.

Bluebox researchers uncovered a variety of weaknesses in the iOS and Android app developed by Mattel partner ToyTalk. The apps are used to connect the doll to a nearby Wi-Fi networks. The researchers also reported vulnerabilities in the remote server used to communicate with the doll.

Insecure baby monitors, hacked dolls are only the leading edges of the Insecure Internet of Things (IIoT).

Dan’s post has the details of the Security-Transmitted-Disease (STD) that can infect Hello Barbie servers and hence the dolls themselves.

When dolls, toys and other devices develop video capabilities, amateur porn will explode on the Insecure Internet of Things (IIoT). With or without the consent of the porn participants.

If you want a secure internet-of-things, the avoid the sieve stacking stacking strategy of current software fixes, which layers broken security software on top of broken software:

present-IT-stack-plus-security

Software security starts from the bottom of your software stack and goes upward.

For all the wailing of software developers about the inability to have perfect software, realize that sql injection attacks were the #1 attack in 2013. That is more than fifteen years after the attack was documented.

Don’t buy into the “we can build perfect software” scam. No one is expecting perfect software, just software that doesn’t have 5+ year old flaws in it.

Is that too much to expect?

Heavy civil penalties for 5+ year old bugs in software might help the software industry remember to avoid such bugs.

December 5, 2015

JAERO: Classic Aero SatCom ACARS signals [Alert: Rendition Trackers]

Filed under: Government,Privacy,Security — Patrick Durusau @ 9:28 pm

JAERO: A program to demodulate and decode Classic Aero SatCom ACARS signals by Jonathan Olds.

From the webpage:

JAERO is a program that demodulates and decodes Classic Aero ACARS (Aircraft Communications Addressing and Reporting System) messages sent from satellites to Aeroplanes (SatCom ACARS) commonly used when Aeroplanes are beyond VHF range. Demodulation is performed using the soundcard. Such signals are typically around 1.5Ghz and can be received with a simple low gain antenna that can be home brewed in a few hours in conjunction with a cheap RTL-SDR dongle.

In the advent of MH370, Classic Aero has become a well-known name. A quick search on the net using “Classic Aero MH370” will produce thousands of results. The Classic Aero signals sent from satellites to the Aeroplanes are what JAERO demodulates and decodes.

I am sure rendition trackers have these and even more sophisticated passive tracking capabilities but I pass this on as a possible starting place for would be civilian surveillance specialists.

Governments are obsessed with surveillance, so much so that civilians need to return the favor, building passive and distributed systems of surveillance that surpass anything governments can obtain.

An interesting hobby to intercept such signals, which are falling in your yard right now, and even more interesting hobby if you capture those signals and share them with other hobbyists. Perhaps even some data science types who can munge the data to bring out interesting insights. Such as on rendition flights.

A couple of tips: Disguise your external receiving equipment to look like a disgruntled satellite TV subscriber (is there any other kind?). Probably a good idea to not discuss monitoring aircraft or other government activities at the local barber shop.

I’m not a conspiracy theorist about government activities but if they didn’t intend someone harm, then why do they keep secrets? If you think about it, none of the many decried data leaks over the last 50 years have resulted in a single American being harmed.

Some of them were embarrassed and probably should have gone to jail (Oliver North) but for all the conjured harms of data leaks, not one has made a whit of difference.

Makes me wonder if secrecy is a means to conceal incompetence and venal criminal wrongdoing.

You?

Big Data Ethics?

Filed under: BigData,Ethics — Patrick Durusau @ 4:10 pm

Ethics are a popular topic in big data and related areas, as I was reminded by Sam Ransbotham’s The Ethics of Wielding an Analytical Hammer.

Here’s a big data ethics problem.

In order to select individuals based on some set of characteristics, habits, etc., we first must define the selection criteria.

Unfortunately, we don’t have a viable profile for terrorists, which explain in part why they can travel under their actual names, with their own identification and not be stopped by the authorities.

So, here’s the ethical question: Is it ethical for contractors and data scientists to offer data mining services to detect terrorists when there is no viable profile for a terrorist?

For all the hand wringing about ethics, basic honesty seems to be in short supply when talking about big data and the search for terrorists.

Yes?

Become a Mini-CIA Today!

Filed under: Government,Privacy,Security — Patrick Durusau @ 3:54 pm

New software watches for license plates, turning you into Little Brother by Cyrus Farivar.

From the post:

We now live in a world where if you have an IP-enabled security camera, you can download some free, open-source software from GitHub and boom—you have a fully functional automated license plate reader (ALPR, or LPR).

How very cool!

I know some privacy advocates may be troubled by this development but think of the old adage:

When guns are outlawed, only outlaws will have guns.

Applying that to surveillance:

When surveillance is outlawed, only outlaws will have surveillance.

Yes?

With the OpenALPR software, neighborhoods that see a lot of police violence can track and alert residents to repeat offenders who are entering the area. Or drug dealers, pimps or other scourges of modern urban areas.

And it would be a counter to suddenly malfunctioning dashboard and body cameras worn by the police.

As I have mentioned before, there are far more citizens than government-based agents. If we start surveillance on them, they will have no place to hide and no where left to run.

Being IP enabled, you could set up a central monitoring station, possibly sharing information with citizens interested in real time “traffic” information.

PS: If you keep the video or scanning results, be sure it is streamed to a heavily encrypted drive.

‘Outsiders’ Crack A 50-Year-Old Math Problem

Filed under: Mathematics,Matrix,Networks — Patrick Durusau @ 3:28 pm

‘Outsiders’ Crack A 50-Year-Old Math Problem by Erica Klarreich.

From the post:

In 2008, Daniel Spielman told his Yale University colleague Gil Kalai about a computer science problem he was working on, concerning how to “sparsify” a network so that it has fewer connections between nodes but still preserves the essential features of the original network.

Network sparsification has applications in data compression and efficient computation, but Spielman’s particular problem suggested something different to Kalai. It seemed connected to the famous Kadison-Singer problem, a question about the foundations of quantum physics that had remained unsolved for almost 50 years.

Over the decades, the Kadison-Singer problem had wormed its way into a dozen distant areas of mathematics and engineering, but no one seemed to be able to crack it. The question “defied the best efforts of some of the most talented mathematicians of the last 50 years,” wrote Peter Casazza and Janet Tremain of the University of Missouri in Columbia, in a 2014 survey article.

As a computer scientist, Spielman knew little of quantum mechanics or the Kadison-Singer problem’s allied mathematical field, called C*-algebras. But when Kalai, whose main institution is the Hebrew University of Jerusalem, described one of the problem’s many equivalent formulations, Spielman realized that he himself might be in the perfect position to solve it. “It seemed so natural, so central to the kinds of things I think about,” he said. “I thought, ‘I’ve got to be able to prove that.’” He guessed that the problem might take him a few weeks.

Instead, it took him five years. In 2013, working with his postdoc Adam Marcus, now at Princeton University, and his graduate student Nikhil Srivastava, now at the University of California, Berkeley, Spielman finally succeeded. Word spread quickly through the mathematics community that one of the paramount problems in C*-algebras and a host of other fields had been solved by three outsiders — computer scientists who had barely a nodding acquaintance with the disciplines at the heart of the problem.

Why all the excitement?


The proof of the Kadison-Singer problem implies that all the constructions in its dozen incarnations can, in principle, be carried out—quantum knowledge can be extended to full quantum systems, networks can be decomposed into electrically similar ones, matrices can be broken into simpler chunks. The proof won’t change what quantum physicists do, but it could have applications in signal processing, since it implies that collections of vectors used to digitize signals can be broken down into smaller frames that can be processed faster. The theorem “has potential to affect some important engineering problems,” Casazza said.

Just so you know, the same people who are saying it will be years before practical results emerge from this breakthrough are the same ones who assumed the answer to this problem was negative. 😉

I’m not saying techniques based on this work will be in JavaScript libraries next year but without trying, they never will be.

Enjoy!

I first saw this in a post by Lars Marius Garshol

December 4, 2015

3 ways to win “Practical Data Science with R”! (Contest ends December 12, 2015 at 11:59pm EST)

Filed under: Contest,Data Science,R — Patrick Durusau @ 5:25 pm

3 ways to win “Practical Data Science with R”!.

Renee is running a contest to give away three copies of “Practical Data Science with R” by Nina Zumel and John Mount!

You must enter on or before December 12, 2015 at 11:59pm EST.

Three ways to win, see Renee’s post for the details!

Witness (activists/change advocates, please read)

Filed under: Journalism,News,Reporting — Patrick Durusau @ 5:15 pm

Witness

From the about page:

WITNESS trains and supports activists and citizens around the world to use video safely, ethically, and effectively to expose human rights abuse and fight for human rights change.

The majority of the world’s population now has a camera in their pocket. People everywhere are turning to video to document and tell stories of abuse. But all too often, they are not filming safely or effectively, and their videos don’t make a difference.

We identify critical situations and teach those affected by them the basics of video production, safe and ethical filming techniques, and advocacy strategies.

A wealth of training materials and videos for anyone who wants to make a difference with their cellphone or video camera.

TV images of the police using water cannon and setting dogs on young people and children in Birmingham, Alabama, were a watershed event for the civil rights movement in the United States.

But that isn’t to promise the technique works every time. Scenes of far worse abuse are common from the Gaza strip and world governments continue to look away.

But you won’t know if the video advocacy will work for you or not unless you try. The Witness site will enable you to do just that.

Many of the techniques described are equally useful for journalists, researchers, etc.

I first saw this at FirstDraftNews as: The Witness Resource Library.

PS: Since I mentioned the Birmingham Children’s Crusade (1963), I should also point out that Malcolm X disagreed with the strategy of sending children into harm’s way. As do I. There are more appropriate ways to confront those of practice violence than sacrificing children for photo ops.

[A] Game-Changing Go Engine (Losing mastery over computers? Hardly.)

Filed under: Artificial Intelligence,Games — Patrick Durusau @ 4:22 pm

How Facebook’s AI Researchers Built a Game-Changing Go Engine

From the post:

One of the last bastions of human mastery over computers is the game of Go—the best human players beat the best Go engines with ease.

That’s largely because of the way Go engines work. These machines search through all possible moves to find the strongest.

While this brute force approach works well in draughts and chess, it does not work well in Go because of the sheer number of possible positions on a board. In draughts, the number of board positions is around 10^20; in chess it is 10^60.

But in Go it is 10^100—that’s significantly more than the number of particles in the universe. Searching through all these is unfeasible even for the most powerful computers.

So in recent years, computer scientists have begun to explore a different approach. Their idea is to find the most powerful next move using a neural network to evaluate the board. That gets around the problem of searching. However, neural networks have yet to match the level of good amateur players or even the best search-based Go engines.

Today, that changes thanks to the work of Yuandong Tian at Facebook AI Research in Menlo Park and Yan Zhu at Rutgers University in New Jersey. These guys have combined a powerful neural network approach with a search-based machine to create a Go engine that plays at an impressively advanced level and has room to improve.

The new approach is based in large part on advances that have been made in neural network-based machine learning in just the last year or two. This is the result of a better understanding of how neural networks work and the availability of larger and better databases to train them.

This is how Tian and Zhu begin. They start with a database of some 250,000 real Go games. They used 220,000 of these as a training database. They used the rest to test the neural network’s ability to predict the next moves that were played in real games.

If you want the full details, check out:

Better Computer Go Player with Neural Network and Long-term Prediction by Yuandong Tian, Yan Zhu.

Abstract:

Competing with top human players in the ancient game of Go has been a long-term goal of artificial intelligence. Go’s high branching factor makes traditional search techniques ineffective, even on leading-edge hardware, and Go’s evaluation function could change drastically with one stone change. Recent works [Maddison et al. (2015); Clark & Storkey (2015)] show that search is not strictly necessary for machine Go players. A pure pattern-matching approach, based on a Deep Convolutional Neural Network (DCNN) that predicts the next move, can perform as well as Monte Carlo Tree Search (MCTS)-based open source Go engines such as Pachi [Baudis & Gailly (2012)] if its search budget is limited. We extend this idea in our bot named darkforest, which relies on a DCNN designed for long-term predictions. Darkforest substantially improves the win rate for pattern-matching approaches against MCTS-based approaches, even with looser search budgets. Against human players, darkforest achieves a stable 1d-2d level on KGS Go Server, estimated from free games against human players. This substantially improves the estimated rankings reported in Clark & Storkey (2015), where DCNN-based bots are estimated at 4k-5k level based on performance against other machine players. Adding MCTS to darkforest creates a much stronger player: with only 1000 rollouts, darkforest+MCTS beats pure darkforest 90% of the time; with 5000 rollouts, our best model plus MCTS beats Pachi with 10,000 rollouts 95.5% of the time.

The author closes with this summary:

This kind of research is still in its early stages, so improvements are likely in the near future. It may be that humans are about to lose their mastery over computers in yet another area.

I may have to go read the article again because the program as described:

  • Did not invent the game of Go or any of its rules.
  • Did not play any of the 220,000 actual Go games used for training.

That is to say that the Game of Go was invented by people and people playing Go supplied the basis for this Go playing computer.

Not to take anything away from the program or these researchers, but humans are hardly about to lose “…mastery over computers in yet another area.”

Humans remain the creators of such games, the source of training data and the measure against which the computer measures itself.

Who do you think is master in such a relationship?*

* Modulo that the DHS wants to make answers from computers to be the basis for violating your civil liberties. But that’s a different type of “mastery” issue.

From God to Google

Filed under: Decision Making — Patrick Durusau @ 1:18 pm

God-Google

I point this out only partially in jest.

Legend has it that the Deity knows the future consequences of present actions.

Google, on the other hand, has the hubris to decide matters that will impact all of us without understanding any future consequences and/or consultation with those to be affected.

Take for example Google’s Loon project.

In brief the Loon project will release a series of balloons that float in the stratosphere to offer global cellphone service, which will enable anyone in the world to connect to the Internet.

Ground-based astronomy is difficult enough with increasing light pollution.

canada-us-mexico

That map is only a portion of the world map at Dark Site Finder

In addition to growing light pollution, astronomers will have to content with random sky junk from Google’s Loon project. Balloons move at the whim of the wind, making it impossible for astronomers to dodge around Google sky junk for their observations.

Don’t imagine that the surveillance potential will be lost on nation states, which will quickly use Google balloons as cover for their own. How are you going to distinguish one balloon from another at the edge of space?

Moreover, despite protests to the contrary, Google’s motivation is fairly clear, the more people with access to the Internet, the more victims of Internet advertising for Google to deliver to its clients.

That the current Internet has adopted an ad-based model is no reason to presume that alternative models on a country by country basis could not be different. Bell telephone survived, nay thrived, for decades without a single ad being delivered by your phone service. Think about that, not a single ad and more reliable service over a wider area than any cellphone provider can boast of today.

Hidden deeper in Google’s agenda is the White West Paternalism that has decided, like being Christian was once upon a time, having access to the Internet is a necessity for everyone. The third world is still paying for missionary intervention in the 17th century and later, must we really repeat that sad episode?

The goal isn’t to benefit people who don’t have access to the Internet but to remake them into more closely fitting our “ideal” of a technology-based society.

The common and infantile belief that technology will bring democracy and other assorted benefits was thoroughly rebutted in 
Why Technology Hasn’t Delivered More Democracy by Thomas Carothers. Democracy and other social benefits are far more complicated, people complicated, than any simple technological “fix” can address.

I think the word I am search for is “hubris.”

Unfortunately, unlike an individual case of “hubris,” when spreading the Internet fails to produce the desired outcomes, it won’t impact only Google, but the rest of humanity as well.

I first saw the Google image in a tweet by Marko A. Rodriguez.

Leaving Pakistan in the Stone Ages

Filed under: Cybersecurity,Government,Privacy,Security — Patrick Durusau @ 10:47 am

BlackBerry gets bounced from Pakistan after saying no to backdoors by John Zorabedian.

From the post:

BlackBerry is saying “no” to government backdoor access to communications on its services and devices, in actions that speak louder than words.

Earlier this week, BlackBerry announced it is shutting down its operations in Pakistan and will leave the country by 30 December, after refusing to provide Pakistan’s government with backdoor access to its customers’ communications.

Marty Beard, BlackBerry’s chief operating officer, wrote on the company’s blog that the Pakistan Telecommunications Authority told mobile phone operators in July that BlackBerry would no longer be allowed to operate in the country for “security reasons.”

Beard said that Pakistan wanted unfettered access to all emails, BBM messages and other Blackberry Enterprise Service (BES) traffic, but the company refused on principle:

Finally, a viable alternative to bombing countries back into the Stone Ages. Just leave them technologically in the Stone Ages. See how their citizens, businesses, crime lords, etc. feel about that!

If Pakistan was demanding backdoors from BlackBerry, you have to wonder what demands have been made on other communication service providers. Yes?

One hopes that such service providers, including those that control the pipes in and out of Pakistan will take stances like that of BlackBerry.

Would be sad to see Pakistan suddenly go dark on the Web but choices do have consequences. Isolated from the modern communications networks used by other countries, government officials will have lots of time for their paranoid fantasies.

Best prepare for a sudden exit of capital and bright minds from Pakistan once it goes dark. Not altogether sure communications should be restored even if the government changes. Let it stay dark as a lesson about governmental over reaching for the rest of the world.

PS: If you want a less extreme lesson first, cut Pakistan off of the Internet for a week, just as a warning to its government.

December 3, 2015

How-To Detect Fake Images in 20 Seconds

Filed under: Journalism,News,Reporting,Verification — Patrick Durusau @ 11:24 am

The video has a good “beat” to it. 😉

Entire video is 42 seconds.

Enjoy!

Beta Testing eFOIA (FBI)

Filed under: Government,Government Data — Patrick Durusau @ 11:11 am

Want to Obtain FBI Records a Little Quicker? Try New eFOIA System

From the post:

The FBI recently began open beta testing of eFOIA, a system that puts Freedom of Information Act (FOIA) requests into a medium more familiar to an ever-increasing segment of the population. This new system allows the public to make online FOIA requests for FBI records and receive the results from a website where they have immediate access to view and download the released information.

Previously, FOIA requests have only been made through regular mail, fax, or e-mail, and all responsive material was sent to the requester through regular mail either in paper or disc format. “The eFOIA system,” says David Hardy, chief of the FBI’s Record/Information Dissemination Section, “is for a new generation that’s not paper-based.” Hardy also notes that the new process should increase FBI efficiency and decrease administrative costs.

The eFOIA system continues in an open beta format to optimize the process for requesters. The Bureau encourages requesters to try eFOIA and to e-mail foipaquestions@ic.fbi.gov with any questions or difficulties encountered while using it. In several months, the FBI plans to move eFOIA into full production mode.

The post gives a list of things you need to know/submit in order to help with beta testing of the eFOIA system.

Why help the FBI?

It’s true, I often chide the FBI for its padding of terrorism statistics by framing the mentally ill and certainly its project management skills are nothing to write home about.

Still, there are men and women in the FBI who do capture real criminals and not just the gullible or people who have offended the recording or movie industries. There are staffers, like the ones behind the eFOIA project, who are trying to do a public service, despite the bad apples in the FBI barrel.

Let’s give them a hand, even though decisions on particular FOIA requests may be quite questionable. Not the fault of the technology or the people who are trying to make it work.

What are you going to submit an FOIA about?

I first saw this in a tweet by Nieman Lab.

Verification around the world: From Kenya to Hong Kong

Filed under: Journalism,News,Reporting,Verification — Patrick Durusau @ 10:47 am

Verification around the world: From Kenya to Hong Kong by Tom Trewinnard.

From the post:

Misinformation around news on the social web is a truly global issue and around the world there are many groups and projects rising to the verification challenge. Here’s a by-no-means-exhaustive run down of a few projects we at Meedan have been watching recently.

Have you seen any great international verification efforts that we’ve missed? Share them with us in the comments here or at @FirstDraftNews on Twitter.

Seven very good examples of verification from a variety of news outlets.

You don’t want to include false information in your topic maps.

How do you verify information for inclusion?

When you debunk rumors or uncover false information, how do you share it with others?

December 2, 2015

‘Not a Math Person’: [Teaching/Communication Strategies]

Filed under: Education,Mathematics — Patrick Durusau @ 3:06 pm

‘Not a Math Person’: How to Remove Obstacles to Learning Math by Katrina Schwartz.

From the post:

Stanford math education professor Jo Boaler spends a lot of time worrying about how math education in the United States traumatizes kids. Recently, a colleague’s 7-year-old came home from school and announced he didn’t like math anymore. His mom asked why and he said, “math is too much answering and not enough learning.”

This story demonstrates how clearly kids understand that unlike their other courses, math is a performative subject, where their job is to come up with answers quickly. Boaler says that if this approach doesn’t change, the U.S. will always have weak math education.

“There’s a widespread myth that some people are math people and some people are not,” Boaler told a group of parents and educators gathered at the 2015 Innovative Learning Conference. “But it turns out there’s no such thing as a math brain.” Unfortunately, many parents, teachers and students believe this myth and it holds them up every day in their math learning.

Intriguing article that suggests the solution to the lack of students in computer science and mathematics may well be to work on changing the attitudes of students…about themselves as computer science or mathematics students.

Something to remember when users are having a hard time grasping your explanation of semantics and/or topic maps.

Oh, another high point in the article, our brains physically swell and shrink:

Neuroscientists now know that the brain has the ability to grow and shrink. This was demonstrated in a study of taxi drivers in London who must memorize all the streets and landmarks in downtown London to earn a license. On average it takes people 12 tries to pass the test. Researchers found that the hippocampus of drivers studying for the test grew tremendously. But when those drivers retired, the brain shrank. Before this, no one knew the brain could grow and shrink like that.

It is only year two of the Human Brain Project and now we know that one neuron can have thousands of synapses and now that the infrastructure of the brain grows and shrinks. Information that wasn’t available at its start.

How do you succeed when the basic structure to be modeled keeps changing?

Perhaps that is why the Human Brain Project has no defined measure of “success”, other than spending all the allotted funds over a ten year period. That I am sure they will accomplish.

Signal Desktop beta! [End-to-end encryption] e3

Filed under: Cybersecurity,Government,Privacy,Security — Patrick Durusau @ 2:36 pm

Signal Desktop beta!

From the post:

Today we’re making the Signal Desktop beta available. Signal Desktop brings the trusted private messaging experience of Signal to the desktop, with a simplicity that allows you to seamlessly continue conversations back and forth between your mobile device and your desktop computer.

Private messaging, now with all ten fingers

As always, everything is end-to-end encrypted and painstakingly engineered in order to keep your communication safe – allowing you to send high-quality private group, text, picture, and video messages for free.

(graphic omitted)

Multiple devices, single identifier

Signal Desktop is a Chrome app which links with your phone, so all incoming and outgoing messages are displayed consistently on all your devices. Your contacts don’t have to guess where to message you, and when you switch devices the conversation you started will already be there.

(graphic omitted)

Android devices only, for now

For the initial Signal desktop beta, only linking to Android devices is supported. Follow us on twitter for updates on when the iOS app supports Signal Desktop.

View source

All of our code is free, open source, and available on GitHub. This allows experts to verify our protocols and our implementations.

Like everything we do at Open Whisper Systems, dedicated development is supported by community donations and grants. Signal Desktop contains no advertisements, and it doesn’t cost anything to use.

Terrorists don’t use encrypted messaging, but that is not a reason for you to avoid end-to-end encryption.

From the How do I help? page:

  • Spread the word – Open WhisperSystems is a collaborative open source project and does not have a dedicated PR department. We rely on our users to help explain the benefits of using our software. Friends don’t let friends send plaintext!
  • Help your fellow users – Join our mailing list and assist new and existing users, or help shape the future of the project by participating in the discussions that are held there. We also appreciate assistance with marking issues as duplicates in our GitHub repos, answering questions that are raised in the form of issues, helping to reproduce bugs while providing additional details, or performing other forms of triage.
  • Contribute code – If you have Android or iOS development experience, please consider helping us tackle some of the open issues in our repositories. Pull requests for new features and functionality are also welcome.
  • Contribute moneyBitHub is our experiment in collaborative funding for open source projects. Donating to our BitHub pool provides an additional incentive for developers to contribute their work and time. You can also donate to Open WhisperSystems via the Freedom of the Press Foundation.

The Open WhisperSystems project is where it is today because of your help and support. Thank you!

To market end-to-end encryption, would e3 be a good logo?

Ex. If e3 is banned, only criminals will have e3.

Twitter Journalism Tips

Filed under: Journalism,News,Reporting,Tweets,Twitter — Patrick Durusau @ 1:49 pm

Twitter Journalism Tips from FirstDraftNews.

Five videos on effective use of Twitter for journalism.

The videos are:

How To Use Twitter Lists For Journalism 2:37

Why I Love Twitter Lists – Sue Llewellyn 1:48

Journalist Guide: How To Use Tweekdeck 1:14

Journalist Tweetdeck Tips – Reuters, George Sargent 1:12

Searching For Geolocated Posts On Twitter 1:26

The times shown are minutes followed by seconds.

Labeled for journalism but anyone searching Twitter, librarians, authors, researchers, even “fans” (shudder), will find useful information in these videos.

If you don’t know FirstDraftNews, you need to get acquainted.

Unicode to LaTeX

Filed under: TeX/LaTeX,Unicode — Patrick Durusau @ 11:26 am

Unicode to LaTeX by John D. Cook.

From the post:

I’ve run across a couple web sites that let you enter a LaTeX symbol and get back its Unicode value. But I didn’t find a site that does the reverse, going from Unicode to LaTeX, so I wrote my own.

Unicode / LaTeX Conversion

If you enter Unicode, it will return LaTeX. If you enter LaTeX, it will return Unicode. It interprets a string starting with “U+” as a Unicode code point, and a string starting with a backslash as a LaTeX command.

I am having trouble visualizing when I would need to go from Unicode to LaTeX but on the off-chance that I find myself in that situation, I wanted to note John’s conversion page.

Knowing my luck, just after this post is pushed off the front page of the blog I will have need of it. 😉

December 1, 2015

New Rule for Software Patches: Don’t Make Things Worse

Filed under: Cybersecurity,Security — Patrick Durusau @ 7:36 pm

Security Advisory: Dell Foundation Services Remote Information Disclosure (II)

From the post:

Dell Foundation Services starts an HTTPd that listens on port 7779. The previous service tag leak was fixed by removing the JSONP API.

However, the webservice in question is still available; it is now a SOAP service, and all methods of that webservice can be accessed, not just the ServiceTag method.

One of the methods accessible is List GetWmiCollection(string wmiQuery) – this returns the results of a given Windows Management Instrumentation (WMI) query, enabling access to information about hardware, installed software, running processes, installed services, accessible hard disks, filesystem metadata (filenames, file size, dates) and more.

Amazing isn’t it?

The post recommends removal of Dell Foundational Services. Same way you cure Adobe Flash security problems.

Are you a debunker or fact-checker? (take the survey, it’s important)

Filed under: Journalism,News,Reporting,Social Media — Patrick Durusau @ 7:17 pm

Are you a debunker or fact-checker? Take this survey to help us understand the issue by Craig Silverman.

From the post:

Major news events like the Paris attacks quickly give birth to false rumors, hoaxes and viral misinformation. But there is a small, growing number of people and organizations who seek to quickly identify, verify or debunk and spread the truth about such misinformation that arises during major news events. As much as possible, they want to stop a false rumor before it spreads.

These real-time debunkers, some of which First Draft has profiled recently, face many challenges. But by sharing such challenges and possible solutions, it is possible to find collective answers to the problem of false information.

The survey below intends to gather tips and tactics from those doing this work in an attempt to identify best practices to be shared.

If you are engaged in this type of work — or have experimented with it in the past — we ask that you please take a moment to complete this quick survey and share your advice.

I don’t know if it is possible for debunking or fact-checking to run faster than rumor and falsehood but that should not keep us from trying.

Definitely take the survey!

Ding-Dong! Flash Is Dying!

Filed under: Cybersecurity,HTML5,Security — Patrick Durusau @ 6:47 pm

Flash, HTML5 and Open Web Standards

From the post:

Adobe has a history of pioneering and advancing industry standards. We embrace standards and, where none exist, we create them.

Flash has played a leading role in bringing new capabilities to the web. From audio and animation, to interactivity and video, Flash has helped push the web forward.

Today, open standards like HTML5 have matured and provide many of the capabilities that Flash ushered in. Our customers have clearly communicated that they would like our creative applications to evolve to support multiple standards and we are committed to doing that. So today we are announcing Animate CC, previously Flash Professional CC, which will be Adobe’s premier web animation tool for developing HTML5 content while continuing to support the creation of Flash content. Adobe Animate CC will be available in early 2016. In addition, Adobe will release an HTML5 video player for desktop browsers, which will complement Adobe’s support for HTML5 on mobile. [Visit the Primetime blog for more information].

I didn’t realize we lacked a standard for web insecurity. Certainly Adobe Flash set the standard for maximum cyberinsecurity. I don’t know that I would brag about it.

Adobe says that new standards aren’t mature in “web gaming and premium video” and so will keep promoting Flash for those use cases. I take that to mean standards geeks and implementation experts need to double down in both of those areas.

The sooner Flash is just an unpleasant memory the more secure we will all be in the present.

I first saw this in a Facebook post by Alex Brown.

Progress on Connecting Votes and Members of Congress (XQuery)

Filed under: Government,Government Data,XQuery — Patrick Durusau @ 5:34 pm

Not nearly to the planned end point but I have corrected a file I generated with XQuery that provides the name-id numbers for members of the House and a link to their websites at house.gov.

It is a rough draft but you can find it at: http://www.durusau.net/publications/name-id-member-website-draft.html.

While I was casting about for the resources for this posting, I had the sinking feeling that I had wasted a lot of time and effort when I found: http://clerk.house.gov/xml/lists/MemberData.xml.

But, if you read that file carefully, what is the one thing it lacks?

A link to every members’s website at “….house.gov.”

Isn’t that interesting?

Of all the things to omit, why that one?

Especially since you can’t auto-generate the website names from the member names. What appear to be older names use just the last name of members. But, that strategy must have fallen pretty quickly when members with the same last names appeared.

The conflicting names and even some non-conflicting names follow a new naming protocol that appears to be firstname+lastname.house.gov.

That will work for a while until the next generation starts inheriting positions in the House.

Anyway, that is as far as I got today but at least it is a useful list for invoking the name-id of members of the House and obtaining their websites.

The next step will be hitting the websites to extract contact information.

Yes, I know that http://clerk.house.gov/xml/lists/MemberData.xml has the “official” contact information, along with their forms for email, etc.

If I wanted to throw my comment into a round file I could do that myself.

No, what I want to extract is their local office data so when they are “back home” meeting with constituents, the average voter has a better chance of being one of those constituents. Not just those who maxed out on campaign donations limits.

« Newer Posts

Powered by WordPress