Another Word For It Patrick Durusau on Topic Maps and Semantic Diversity

October 6, 2015

Promises Of Stronger Protections From A Habitual Liar?

Filed under: Cybersecurity,Government,Privacy,Security — Patrick Durusau @ 7:41 pm

Europe’s highest court strikes down Safe Harbor data sharing between EU, US by Sebastian Anthony.

From the post:

Europe’s top court, the Court of Justice of the European Union (CJEU), has struck down the 15-year-old Safe Harbour agreement that allowed the free flow of information between the US and EU. The most significant repercussion of this ruling is that American companies, such as Facebook, Google, and Twitter, may not be allowed to send user data from Europe back to the US.

The full text of the decision: decision (link to the full text).

A repeated theme in discussion of this decision is the need for stronger promises by the U.S. to protect European privacy rights.

I’ll be the first to admit that I don’t follow some segments of the news very closely but surely most people have heard of Edward Snowden. Yes?

I won’t recite the history of his disclosures here but suffice it to say that his revelations establish beyond any doubt that the United States government has systematically disobeyed it own laws and the laws of other countries in surveillance and other areas. If that weren’t bad enough, the U.S. government has repeated lied to the people it governs and other countries.

Let’s assume that the United States government agrees to very strong provisions for guarding the privacy of EU citizens. On what basis would you trust such a promise? A government willing to break it own laws, to lie to its own people, certainly will have no qualms lying to other countries.

In litigation that challenges any future agreement on the transfer of user data from Europe to the United States, the Court of Justice of the European Union (CJEU) should take judicial notice that the United States is in fact a habitual liar and its word counts for nothing in its proceedings.

I don’t know how long it will take the United States to regain credibility in international courts but it has fully and well earned the designation “habitual liar” in present proceedings.

Getting Rid of “Get Windows 10!” (Public Service Announcement)

Filed under: Cybersecurity,Microsoft — Patrick Durusau @ 10:13 am

There is a difference between commercial software and nagware. Or, there was once upon a time. To promote Window 10, a Microsoft ad has taken up residency in the system tray of Windows 7 and Windows 8 users.

To revert to a non-nagware version of Windows 7 or Windows 8, see: What is the “Get Windows 10” Tray Item and How Do You Remove It?

Bob Ducharme reports this worked for him.

I haven’t taken the Windows 10 plunge (onto a VM) but then I encountered this language in the Window 10 EULA:

Updates. The softwareperiodically checks for system and app updates, and downloads and installs them for you. You may obtain updates only from Microsoft or authorized sources, and Microsoft may need to update your system to provide you with those updates. By accepting this agreement, you agree to receive these types of automatic updates without any additional notice.
[emphasis added to last sentence]

If you are using Windows 10 to read email and surf the web, that may be ok.

If you are building mission-critical applications that rely on the stability of Windows system calls, that’s insane.

Ask you IT department about MS “updates” that have toasted applications in the past.

If Windows 10 becomes the dog that whatever came right after Windows XP did (I can’t even remember its name), perhaps Microsoft will adopt saner update policy for Windows (whatever).

October 5, 2015

International Hysteria Over American Gun Violence

Filed under: Data Mining,News,Text Mining — Patrick Durusau @ 7:56 pm

Australia’s call for a boycott on U.S. travel until gun-reform is passed may be the high point of the international hysteria over gun violence in the United States. Or it may not be. Hard to say at this point.

Social media has been flooded with hand wringing over the loss of “innocent” lives, etc., you know the drill.

The victims in Oregon were no doubt “innocent,” but innocence alone isn’t the criteria by which “mass murder” is judged.

At least not according to both the United States government, other Western governments and their affiliated news organizations.

Take the Los Angeles Times for example, which has an updated list of mass shootings, 1984 – 2015.

Or the breathless prose of The Chicagoist in Chicago Dominates The U.S. In Mass Shootings Count.

Based on data compiled by the crowd-sourced Mass Shooting Tracker site, the Guardian discovered that there were 994 mass shootings—defined as an incident in which four or more people are shot—in 1,004 days since Jan. 1, 2013. The Oregon shooting happened on the 274th day of 2015 and was the 294th mass shooting of the year in the U.S.

Some 294 mass shootings since January 1, 2015 in the U.S.?

Chump change my friend, chump change.

No disrespect to the innocent dead, wounded or their grieving families, but as I said, “innocence isn’t the criteria for judging mass violence. Not by Western governments, not by the Western press.

You will have to do a little data mining to come to that conclusion but if you have the time, follow along.

First, of course, we have to find acts of violence with no warning to its innocent victims who were just going about their lives. At least until pain and death came raining out of the sky.

Let’s start with Operation Inherent Resolve: Targeted Operations Against ISIL Terrorists.

If you select a country name, your options are Syria and Iraq, a pop-up will display the latest news briefing on “Airstrikes in Iraq and Syria.” Under the current summary, you will see “View Information on Previous Airstrikes.”

Selecting “View Information on Previous Airstrikes” will give you a very long drop down page with previous air strike reports. It doesn’t list human casualties or the number of bombs dropped, but it does recite the number of airstrikes.

Capture that information down to January 1, 2015 and save it to a text file. I have already captured it and you can download us-airstrikes-iraq-syria.txt.

You will notice that the file has text other than the air strikes, but air strikes are reported in a common format:

 - Near Al Hasakah, three strikes struck three separate ISIL tactical units 
   and destroyed three ISIL structures, two ISIL fighting positions, and an 
   ISIL motorcycle.
 - Near Ar Raqqah, one strike struck an ISIL tactical unit.
 - Near Mar’a, one strike destroyed an ISIL excavator.
 - Near Washiyah, one strike damaged an ISIL excavator.

Your first task is to extract just the lines that start with: “- Near” and save them to a file.

I used: grep '\- Near' us-airstrikes-iraq-syria.txt > us-airstrikes-iraq-syria-strikes.txt

Since I now have all the lines with airstrike count data, how do I add up all the numbers?

I am sure there is an XQuery solution but its throw-away data , so I took the easy way out:

grep 'one airstrike' us-airstrikes-iraq-syria-strikes.txt | wc -l

Which gave me a count of all the lines with “one airstrike,” or 629 if you are interested.

Just work your way up through “ten airstrikes” and after that, nothing but zeroes. Multiple the number of lines times the number in the search expression and you have the number of airstrikes for that number. One I found was 132 for “four airstrikes,” so that was 528 airstrikes for that number.

Oh, I forgot to mention, some of the reports don’t use names for numbers but digits. Yeah, inconsistent data.

The dirty answer to that was:

grep '[0-9] airstrikes' us-airstrikes-iraq-syria-strikes.txt > us-airstrikes-iraq-syria-strikes-digits.txt

The “[0-9]” detects any digit, between zero and nine. Could have made it a two-digit number but any two-digit number starts with one digit so why bother?

Anyway, that found another 305 airstrikes that were reported in digits.

Ah, total number of airstrikes, not bombs but airstrikes since January 1, 2015?

4,207 airstrikes as of today.

That’s four thousand, two hundred and seven (minimum, more than one bomb per airstrike), times that innocent civilians may have been murdered or at least terrorized by violence falling out of the sky.

Those 4,207 events were not the work of marginally functional, disturbed or troubled individuals. No, those events were orchestrated by highly trained, competent personnel, backed by the largest military machine on the planet and a correspondingly large military industrial complex.

I puzzle over the international hysteria over American gun violence when the acts are random, unpredictable and departures from the norm. Think of all the people with access to guns in the United States who didn’t go on violent rampages.

The other puzzlement is that the crude data mining I demonstrated above establishes the practice of violence against innocents is a long standing and respected international practice.

Why stress over 294 mass shootings in the U.S. when 4,207 airstrikes in 2015 have killed or endangered equally innocent civilians who are non-U.S. citizens?

What is fair for citizens of one country should be fair for citizens of every country. The international community seems to be rather selective when applying that principle.

October 4, 2015

8,400 NASA Apollo Moon Mission Photos

Filed under: Science — Patrick Durusau @ 9:18 pm

Over 8,400 NASA Apollo moon mission photos just landed online, in high-resolution by Xeni Jardin.

From the post:

Space fans, rejoice: today, just about every image captured by Apollo astronauts on lunar missions is now on the Project Apollo Archive Flickr account. There are some 8,400 photographs in all at a resolution of 1800 dpi, and they’re sorted by the roll of film they were on.

The Project Apollo Archive is also on Facebook. They’ll be showcasing new renderings of some of the best imagery, and other rare images including Apollo 11 training photos.

The Apollo astronauts were sent to the moon with Hasselblad cameras, and the resulting prints have been painstakingly restored for contemporary high-resolution screens for this wonderful archival project.

Long live space.

Tear yourself away from news feeds humming with the latest non-events accompanied by screaming headlines.

The U.S. space program did not unify everyone and there were a multitude of problems (still present) on the ground.

But, it represents what can be achieved by a government that isn’t trying to avoid blame for random and unpreventable acts.

October 3, 2015

Computation + Journalism Symposium 2015

Filed under: Journalism,News,Reporting — Patrick Durusau @ 8:02 pm

Computation + Journalism Symposium 2015

From the webpage:

Data and computation drive our world, often without sufficient critical assessment or accountability. Journalism is adapting responsibly—finding and creating new kinds of stories that respond directly to our new societal condition. Join us for a two-day conference exploring the interface between journalism and computing.

Papers are up! Papers are up!

http://cj2015.brown.columbia.edu/papers.html

Many excellent papers but one caught my eye in particular:

DeScipher: A Text Simplification Tool for Science Journalism, Yea Seul Kim, Jessica Hullman and Eytan Adar.

High on my reading list after spending a day with “almost” explanations in technical documentation.

This could be very useful for anyone authoring useful technical documentation, not to mention writing for the general public.

October 2, 2015

Workflow for R & Shakespeare

Filed under: Literature,R,Text Corpus,Text Mining — Patrick Durusau @ 2:00 pm

A new data processing workflow for R: dplyr, magrittr, tidyr, ggplot2

From the post:

Over the last year I have changed my data processing and manipulation workflow in R dramatically. Thanks to some great new packages like dplyr, tidyr and magrittr (as well as the less-new ggplot2) I've been able to streamline code and speed up processing. Up until 2014, I had used essentially the same R workflow (aggregate, merge, apply/tapply, reshape etc) for more than 10 years. I have added a few improvements over the years in the form of functions in packages doBy, reshape2 and plyr and I also flirted with the package data.table (which I found to be much faster for big datasets but the syntax made it difficult to work with) — but the basic flow has remained remarkably similar. Until now…

Given how much I've enjoyed the speed and clarity of the new workflow, I thought I would share a quick demonstration.

In this example, I am going to grab data from a sample SQL database provided by Google via Google BigQuery and then give examples of manipulation using dplyr, magrittr and tidyr (and ggplot2 for visualization).

This is a great introduction to a work flow in R that you can generalize for your own purposes.

Word counts won’t impress your English professor but you will have a base for deeper analysis of Shakespeare.

I first saw this in a tweet by Christophe Lalanne.

Emacs Mini Manual, etc.

Filed under: Editor — Patrick Durusau @ 1:48 pm

Emacs Mini Manual, etc.

From the webpage:

Very strong resources on Emacs for programmers.

The animated graphics are a real treat.

I first saw this in a tweet by Christophe Lalanne.

Debugging XQuery Advice

Filed under: XQuery — Patrick Durusau @ 6:02 am

If you are ever called upon to diagnose or repair network connectivity issues, you know that the first thing to check is the network cable.

Well, much to my chagrin, there is a similar principle to follow when debugging XQuery statements.

If you type an element name incorrectly, you may not get an error from the query and it will happily complete, sans your expected content.

To broaden that a bit, the first thing to check, outside of reported syntax and type errors, are your XPath expressions, including the spelling of element names.

Especially for no errors but also not the expected result cases.

Thus ends the XQuery lesson for the day.

October 1, 2015

Stagefright Bug 2.0 [/bettertargets.txt ?]

Filed under: Cybersecurity,Security — Patrick Durusau @ 6:40 pm

Stagefright Bug 2.0 – One Billion Android SmartPhones Vulnerable to Hacking by Mohit Kumar.

From the post:

Attention Android users!

More than 1 Billion Android devices are vulnerable to hackers once again – Thanks to newly disclosed two new Android Stagefright vulnerabilities.

Yes, Android Stagefright bug is Back…

…and this time, the flaw allows an attacker to hack Android smartphones just by tricking users into visiting a website that contains a malicious multimedia file, either MP3 or MP4.

For all the talk about better software, better security procedures, etc., nothing seems to be really cost-effective at stopping hacking.

Instead of putting our limited fingers into the increasing number of cyber vulnerabilities, may I suggest we take a page from the history of /robots.txt?

In addition to your robots.txt file at the root of your web server, create a bettertargets.txt file also at the root of your file system.

List other organizations, government agencies, etc. that have more valuable information assets than you and any information you have that could be used to breach those sites.

Hackers should appreciate the assist and the higher quality assets they can obtain at other sites. At the least it will get them to move away from your machine, which is the point of cybersecurity, at least from a personal point of view.

As a suggested format, a plain tab-delimited text file where each line begins with the domain-name tab IP-address tab assets-(summary of information assets) tab vulnerability-(description of vulnerability).

Suggestions for enhancements?

Is the term “tease” still in fashion?

Filed under: Advertising,Books — Patrick Durusau @ 4:32 pm

I ask if “tease” is still in fashion (or its more sexist equivalent) because I keep running across partial O’Reilly publications that are touted as “free,” but are in reality, just extended ads for forthcoming books.

A case in point is “Transforms in CSS” which isn’t really a book but an excerpt from the forth edition of CSS: The Definitive Guide.

Forty page book?

Social media with light up with posts and reposts about this “free” title.

Save your time and disk space. If anything, get a preview copy of the forth edition of CSS: The Definitive Guide when it is available.

Make no mistake, I like O’Reilly publications and I am presently reading what I suspect is the best O’Reilly title in a number of years, XQuery by Priscilla Walmsley.

O’Reilly shouldn’t waste bandwidth with disconnected excerpts for its titles.

Federal Cybersecurity: More Holes Than Swiss Cheese

Filed under: Cybersecurity,Government,Security — Patrick Durusau @ 3:58 pm

Agencies Need to Correct Weaknesses and Fully Implement Security Programs GAO-15-714: Published: Sep 29, 2015.

From the webpage:

Persistent weaknesses at 24 federal agencies illustrate the challenges they face in effectively applying information security policies and practices. Most agencies continue to have weaknesses in (1) limiting, preventing, and detecting inappropriate access to computer resources; (2) managing the configuration of software and hardware; (3) segregating duties to ensure that a single individual does not have control over all key aspects of a computer-related operation; (4) planning for continuity of operations in the event of a disaster or disruption; and (5) implementing agency-wide security management programs that are critical to identifying control deficiencies, resolving problems, and managing risks on an ongoing basis (see fig.). These deficiencies place critical information and information systems used to support the operations, assets, and personnel of federal agencies at risk, and can impair agencies’ efforts to fully implement effective information security programs. In prior reports, GAO and inspectors general have made hundreds of recommendations to agencies to address deficiencies in their information security controls and weaknesses in their programs, but many of these recommendations remain unimplemented.

Can you guess why “…may of these recommendations remain unimplemented?

The first and foremost reason is that disregarding a recommendation by the GAO or inspectors general has no consequences, none.

Can you imagine being in charge of maintaining your corporate firewall and when it is breached telling your boss, “yeah, I know you said to fix the old one but I got busy and just never did it.”

What do you think the consequences for you personally would be? (You have only one guess.)

It doesn’t appear to work like that at federal agencies. The same people make the same mistakes, over and over again, with no consequences whatsoever.

The only way to change the current cybersecurity state of federal agencies is to provide consequences for failure to improve.

The GAO and inspectors general should be given day to day control over agency spending and personnel decisions as they relate to cybersecurity priorities. And empowered to hire and fire staff as they see fit.

Any other remedy is a recipe for federal security that barely test script kiddies, much less more serious international opponents.

« Newer Posts

Powered by WordPress