Another Word For It Patrick Durusau on Topic Maps and Semantic Diversity

November 24, 2015

What’s the significance of 0.05 significance?

Filed under: Mathematics,Statistics — Patrick Durusau @ 11:08 am

What’s the significance of 0.05 significance? by Carl Anderson.

From the post:

Why do we tend to use a statistical significance level of 0.05? When I teach statistics or mentor colleagues brushing up, I often get the sense that a statistical significance level of α = 0.05 is viewed as some hard and fast threshold, a publishable / not publishable step function. I’ve seen grad students finish up an empirical experiment and groan to find that p = 0.052. Depressed, they head for the pub. I’ve seen the same grad students extend their experiment just long enough for statistical variation to swing in their favor to obtain p = 0.049. Happy, they head for the pub.

Clearly, 0.05 is not the only significance level used. 0.1, 0.01 and some smaller values are common too. This is partly related to field. In my experience, the ecological literature and other fields that are often plagued by small sample sizes are more likely to use 0.1. Engineering and manufacturing where larger samples are easier to obtain tend to use 0.01. Most people in most fields, however, use 0.05. It is indeed the default value in most statistical software applications.

This “standard” 0.05 level is typically associated with Sir R. A. Fisher, a brilliant biologist and statistician that pioneered many areas of statistics, including ANOVA and experimental design. However, the true origins make for a much richer story.

One of the best history/explanations of 0.05 significance I have ever read. Highly recommended!

In part because in the retelling of this story Carl includes references that will allow you to trace the story in even greater detail.

What is dogma today, 0.05 significance, started as a convention among scientists, without theory, without empirical proof, without any of gate keepers associated with scientific publishing of today.

Over time 0.05 significance has proved its utility. The question for you is what other dogmas of today rely on the chance practices of yesteryear?

I first saw this in a tweet by Kirk Borne.

October 13, 2015

How to teach gerrymandering…

Filed under: Government,Mathematics — Patrick Durusau @ 2:26 pm

How to teach gerrymandering and its many subtle, hard problems by Cory Doctorow.

From the post:

Ben Kraft teaches a unit on gerrymandering — rigging electoral districts to ensure that one party always wins — to high school kids in his open MIT Educational Studies Program course. As he describes the problem and his teaching methodology, I learned that district-boundaries have a lot more subtlety and complexity than I’d imagined at first, and that there are some really chewy math and computer science problems lurking in there.

Kraft’s pedagogy is lively and timely and extremely relevant. It builds from a quick set of theoretical exercises and then straight into contemporary, real live issues that matter to every person in every democracy in the world. This would be a great unit to adapt for any high school civics course — you could probably teach it in middle school, too.

Certainly timely considering that congressional elections are ahead (in the United States) in 2016.

Also a reminder that in real life situations, mathematics, algorithms, computers, etc., are never neutral.

The choices you make determine who will serve and who will eat.

It was ever thus and those who pretend otherwise are trying to hide their hand on the scale.

September 27, 2015

What Does Probability Mean in Your Profession? [Divergences in Meaning]

Filed under: Mathematics,Probability,Subject Identity — Patrick Durusau @ 9:39 pm

What Does Probability Mean in Your Profession? by Ben Orlin.

Impressive drawings that illustrate the divergence in meaning of “probability” for various professions.

I’m not sold on the “actual meaning” drawing because if everyone in a discipline understands “probability” to mean something else, on what basis can you argue for the “actual meaning?”

If I am reading a paper by someone who subscribes to a different meaning than your claimed “actual” one, then I am going to reach erroneous conclusions about their paper. Yes?

That is in order to understand a paper I have to understand the words as they are being used by the author. Yes?

If I understand “democracy and freedom” to mean “serves the interest of U.S.-based multinational corporations,” then calls for “democracy and freedom” in other countries isn’t going to impress me all that much.

Enjoy the drawings!

May 31, 2015

Mathematics: Best Sets of Lecture Notes and Articles

Filed under: Mathematics — Patrick Durusau @ 9:47 am

Mathematics: Best Sets of Lecture Notes and Articles by Alex Youcis.

From the post:

Let me start by apologizing if there is another thread on math.se that subsumes this.

I was updating my answer to the question here during which I made the claim that “I spend a lot of time sifting through books to find [the best source]”. It strikes me now that while I love books (I really do) I often find that I learn best from sets of lecture notes and short articles. There are three particular reasons that make me feel this way.

1.Lecture notes and articles often times take on a very delightful informal approach. They generally take time to bring to the reader’s attention some interesting side fact that would normally be left out of a standard textbook (lest it be too big). Lecture notes and articles are where one generally picks up on historical context, overarching themes (the “birds eye view”), and neat interrelations between subjects.

2.It is the informality that often allows writers of lecture notes or expository articles to mention some “trivial fact” that every textbook leaves out. Whenever I have one of those moments where a definition just doesn’t make sense, or a theorem just doesn’t seem right it’s invariably a set of lecture notes that sets everything straight for me. People tend to be more honest in lecture notes, to admit that a certain definition or idea confused them when they first learned it, and to take the time to help you understand what finally enabled them to make the jump.

3.Often times books are very outdated. It takes a long time to write a book, to polish it to the point where it is ready for publication. Notes often times are closer to the heart of research, closer to how things are learned in the modern sense.

It is because of reasons like this that I find myself more and more carrying around a big thick manila folder full of stapled together articles and why I keep making trips to Staples to get the latest set of notes bound.

So, if anyone knows of any set of lecture notes, or any expository articles that fit the above criteria, please do share!

I’ll start:

Fascinating collections of lecture notes and articles, in no particular order with duplication virtually guaranteed. Still, as a browsing resource or if you want to clean it up for others, it is a great resource.

Enjoy!

GNU Octave 4.0

Filed under: Mathematics — Patrick Durusau @ 7:38 am

GNU Octave 4.0

From the webpage:

GNU Octave is a high-level interpreted language, primarily intended for numerical computations. It provides capabilities for the numerical solution of linear and nonlinear problems, and for performing other numerical experiments. It also provides extensive graphics capabilities for data visualization and manipulation. Octave is normally used through its interactive command line interface, but it can also be used to write non-interactive programs. The Octave language is quite similar to Matlab so that most programs are easily portable.

Version 4.0.0 has been released and is now available for download. Octave 4.0 is a major new release with many new features, including a graphical user interface, support for classdef object-oriented programming, better compatibility with Matlab, and many new and improved functions.

An official Windows binary installer is also available from ftp://ftp.gnu.org/gnu/octave/windows/octave-4.0.0_0-installer.exe

A list of important user-visible changes is availble at http://octave.org/NEWS-4.0.html, by selecting the Release Notes item in the News menu of the GUI, or by typing news at the Octave command prompt.

In terms of documentation:

Reference Manual

Octave is fully documented by a comprehensive 800 page manual.

The on-line HTML and PDF versions of the manual are generated directly from the Texinfo source files that are distributed along with every copy of the Octave source code. The complete text of the manual is also available at the Octave prompt using the doc command.

A printed version of the Octave manual may be ordered from Network Theory, Ltd.. Any money raised from the sale of this book will support the development of free software. For each copy sold $1 will be donated to the GNU Octave Development Fund.

May 19, 2015

The Applications of Probability to Cryptography

Filed under: Cryptography,Mathematics — Patrick Durusau @ 1:25 pm

The Applications of Probability to Cryptography by Alan M. Turing.

From the copyright page:

The underlying manuscript is held by the National Archives in the UK and can be accessed at www.nationalarchives.gov.uk using reference number HW 25/37. Readers are encouraged to obtain a copy.

The original work was under Crown copyright, which has now expired, and the work is now in the public domain.

You can go directly to the record page: http://discovery.nationalarchives.gov.uk/details/r/C11510465.

To get a useful image, you need to add the item to your basket for £3.30.

The manuscript is a mixture of typed text with inserted mathematical expressions added by hand (along with other notes and corrections). This is a typeset version that attempts to capture the original manuscript.

Another recently declassified Turning paper (typeset): The Statistics of Repetition.

Important reads. Turing would appreciate the need to exclude government from our day to day lives.

May 16, 2015

The tensor renaissance in data science

Filed under: Data Science,Mathematics,Tensors — Patrick Durusau @ 8:02 pm

The tensor renaissance in data science by Ben Lorica.

From the post:

After sitting in on UC Irvine Professor Anima Anandkumar’s Strata + Hadoop World 2015 in San Jose presentation, I wrote a post urging the data community to build tensor decomposition libraries for data science. The feedback I’ve gotten from readers has been extremely positive. During the latest episode of the O’Reilly Data Show Podcast, I sat down with Anandkumar to talk about tensor decomposition, machine learning, and the data science program at UC Irvine.

Modeling higher-order relationships

The natural question is: why use tensors when (large) matrices can already be challenging to work with? Proponents are quick to point out that tensors can model more complex relationships. Anandkumar explains:

Tensors are higher order generalizations of matrices. While matrices are two-dimensional arrays consisting of rows and columns, tensors are now multi-dimensional arrays. … For instance, you can picture tensors as a three-dimensional cube. In fact, I have here on my desk a Rubik’s Cube, and sometimes I use it to get a better understanding when I think about tensors. … One of the biggest use of tensors is for representing higher order relationships. … If you want to only represent pair-wise relationships, say co-occurrence of every pair of words in a set of documents, then a matrix suffices. On the other hand, if you want to learn the probability of a range of triplets of words, then we need a tensor to record such relationships. These kinds of higher order relationships are not only important for text, but also, say, for social network analysis. You want to learn not only about who is immediate friends with whom, but, say, who is friends of friends of friends of someone, and so on. Tensors, as a whole, can represent much richer data structures than matrices.

The passage:

…who is friends of friends of friends of someone, and so on. Tensors, as a whole, can represent much richer data structures than matrices.

caught my attention.

The same could be said about other data structures, such as graphs.

I mention graphs because data representations carry assumptions and limitations that aren’t labeled for casual users. Such as directed acyclic graphs not supporting the representation of husband-wife relationships.

BTW, the Wikipedia entry on tensors has this introduction to defining tensor:

There are several approaches to defining tensors. Although seemingly different, the approaches just describe the same geometric concept using different languages and at different levels of abstraction.

Wonder if there is a mapping between the components of the different approaches?

Suggestions of other tensor resources appreciated!

April 24, 2015

Mathematicians Reduce Big Data Using Ideas from Quantum Theory

Filed under: Data Reduction,Mathematics,Physics,Quantum — Patrick Durusau @ 8:20 pm

Mathematicians Reduce Big Data Using Ideas from Quantum Theory by M. De Domenico, V. Nicosia, A. Arenas, V. Latora.

From the post:

A new technique of visualizing the complicated relationships between anything from Facebook users to proteins in a cell provides a simpler and cheaper method of making sense of large volumes of data.

Analyzing the large volumes of data gathered by modern businesses and public services is problematic. Traditionally, relationships between the different parts of a network have been represented as simple links, regardless of how many ways they can actually interact, potentially loosing precious information. Only recently a more general framework has been proposed to represent social, technological and biological systems as multilayer networks, piles of ‘layers’ with each one representing a different type of interaction. This approach allows a more comprehensive description of different real-world systems, from transportation networks to societies, but has the drawback of requiring more complex techniques for data analysis and representation.

A new method, developed by mathematicians at Queen Mary University of London (QMUL), and researchers at Universitat Rovira e Virgili in Tarragona (Spain), borrows from quantum mechanics’ well tested techniques for understanding the difference between two quantum states, and applies them to understanding which relationships in a system are similar enough to be considered redundant. This can drastically reduce the amount of information that has to be displayed and analyzed separately and make it easier to understand.

The new method also reduces computing power needed to process large amounts of multidimensional relational data by providing a simple technique of cutting down redundant layers of information, reducing the amount of data to be processed.

The researchers applied their method to several large publicly available data sets about the genetic interactions in a variety of animals, a terrorist network, scientific collaboration systems, worldwide food import-export networks, continental airline networks and the London Underground. It could also be used by businesses trying to more readily understand the interactions between their different locations or departments, by policymakers understanding how citizens use services or anywhere that there are large numbers of different interactions between things.

You can hop over to Nature, Structural reducibility of multilayer networks, where if you don’t have an institutional subscription:

ReadCube: $4.99 Rent, $9.99 to buy, or Purchase a PDF for $32.00.

Let me save you some money and suggest you look at:

Layer aggregation and reducibility of multilayer interconnected networks

Abstract:

Many complex systems can be represented as networks composed by distinct layers, interacting and depending on each others. For example, in biology, a good description of the full protein-protein interactome requires, for some organisms, up to seven distinct network layers, with thousands of protein-protein interactions each. A fundamental open question is then how much information is really necessary to accurately represent the structure of a multilayer complex system, and if and when some of the layers can indeed be aggregated. Here we introduce a method, based on information theory, to reduce the number of layers in multilayer networks, while minimizing information loss. We validate our approach on a set of synthetic benchmarks, and prove its applicability to an extended data set of protein-genetic interactions, showing cases where a strong reduction is possible and cases where it is not. Using this method we can describe complex systems with an optimal trade–off between accuracy and complexity.

Both articles have four (4) illustrations. Same four (4) authors. The difference being the second one is at http://arxiv.org. Oh, and it is free for downloading.

I remain concerned by the focus on reducing the complexity of data to fit current algorithms and processing models. That said, there is no denying that such reduction methods have proven to be useful.

The authors neatly summarize my concerns with this outline of their procedure:

The whole procedure proposed here is sketched in Fig. 1 and can be summarised as follows: i) compute the quantum Jensen-Shannon distance matrix between all pairs of layers; ii) perform hierarchical clustering of layers using such a distance matrix and use the relative change of Von Neumann entropy as the quality function for the resulting partition; iii) finally, choose the partition which maximises the relative information gain.

With my corresponding concerns:

i) The quantum Jensen-Shannon distance matrix presumes a metric distance for its operations, which may or may not reflect the semantics of the layers (or than by simplifying assumption).

ii) The relative change of Von Neumann entropy is a difference measurement based upon an assumed metric, which may or not represent the underlying semantics of the relationships between layers.

iii) The process concludes by maximizing a difference measurement based upon an assigned metric, which has been assigned to the different layers.

Maximizing a difference, based on an entropy calculation, which is itself based on an assigned metric doesn’t fill me with confidence.

I don’t doubt that the technique “works,” but doesn’t that depend upon what you think is being measured?

A question for the weekend: Do you think this is similar to the questions about dividing continuous variables into discrete quantities?

Ordinary Least Squares Regression: Explained Visually

Filed under: Mathematics,Visualization — Patrick Durusau @ 2:55 pm

Ordinary Least Squares Regression: Explained Visually by Victor Powell and Lewis Lehe.

From the post:

Statistical regression is basically a way to predict unknown quantities from a batch of existing data. For example, suppose we start out knowing the height and hand size of a bunch of individuals in a “sample population,” and that we want to figure out a way to predict hand size from height for individuals not in the sample. By applying OLS, we’ll get an equation that takes hand size—the ‘independent’ variable—as an input, and gives height—the ‘dependent’ variable—as an output.

Below, OLS is done behind-the-scenes to produce the regression equation. The constants in the regression—called ‘betas’—are what OLS spits out. Here, beta_1 is an intercept; it tells what height would be even for a hand size of zero. And beta_2 is the coefficient on hand size; it tells how much taller we should expect someone to be for a given increment in their hand size. Drag the sample data to see the betas change.

[interactive graphic omitted]

At some point, you probably asked your parents, “Where do betas come from?” Let’s raise the curtain on how OLS finds its betas.

Error is the difference between prediction and reality: the vertical distance between a real data point and the regression line. OLS is concerned with the squares of the errors. It tries to find the line going through the sample data that minimizes the sum of the squared errors. Below, the squared errors are represented as squares, and your job is to choose betas (the slope and intercept of the regression line) so that the total area of all the squares (the sum of the squared errors) is as small as possible. That’s OLS!

The post includes a visual explanation of ordinary least squares regression up to 2 independent variables (3-D).

Height wasn’t the correlation I heard with hand size but Visually Explained is a family friendly blog. And to be honest, I got my information from another teenager (at the time), so my information source is suspect.

March 18, 2015

Open Source Tensor Libraries For Data Science

Filed under: Data Science,Mathematics,Open Source,Programming — Patrick Durusau @ 5:20 pm

Let’s build open source tensor libraries for data science by Ben Lorica.

From the post:

Data scientists frequently find themselves dealing with high-dimensional feature spaces. As an example, text mining usually involves vocabularies comprised of 10,000+ different words. Many analytic problems involve linear algebra, particularly 2D matrix factorization techniques, for which several open source implementations are available. Anyone working on implementing machine learning algorithms ends up needing a good library for matrix analysis and operations.

But why stop at 2D representations? In a recent Strata + Hadoop World San Jose presentation, UC Irvine professor Anima Anandkumar described how techniques developed for higher-dimensional arrays can be applied to machine learning. Tensors are generalizations of matrices that let you look beyond pairwise relationships to higher-dimensional models (a matrix is a second-order tensor). For instance, one can examine patterns between any three (or more) dimensions in data sets. In a text mining application, this leads to models that incorporate the co-occurrence of three or more words, and in social networks, you can use tensors to encode arbitrary degrees of influence (e.g., “friend of friend of friend” of a user).

In case you are interested, Wikipedia has a list of software packages for tensor analaysis.

Not mentioned by Wikipedia: Facebook open sourcing TH++ last year, a library for tensor analysis. Along with fblualibz, which includes a bridge between Python and Lua (for running tensor analysis).

Uni10 wasn’t mentioned by Wikipedia either.

Good starting place: Big Tensor Mining, Carnegie Mellon Database Group.

Suggest you join an existing effort before you start duplicating existing work.

February 26, 2015

Category Theory – Reading List

Filed under: Category Theory,Mathematics — Patrick Durusau @ 4:35 pm

Category Theory – Reading List by Peter Smith.

Notes along with pointers to other materials.

About Peter Smith:

These pages are by me, Peter Smith. I retired in 2011 from the Faculty of Philosophy at Cambridge. It was my greatest good fortune to have secure, decently paid, university posts for forty years in leisurely times, with a very great deal of freedom to follow my interests wherever they led. Like many of my generation, I am sure I didn’t at the time really appreciate just how lucky I and my contemporaries were. Some of the more student-orientated areas of this site, then, such as the Teach Yourself Logic Guide, constitute my small but heartfelt effort to give something back by way of thanks.

There is much to explore at Peter’s site beside his notes on category theory.

February 23, 2015

Category theory for beginners

Filed under: Category Theory,Education,Language,Mathematics — Patrick Durusau @ 2:09 pm

Category theory for beginners by Ken Scrambler

From the post:

Explains the basic concepts of Category Theory, useful terminology to help understand the literature, and why it’s so relevant to software engineering.

Some two hundred and nine (209) slides, ending with pointers to other resources.

I would have dearly loved to see the presentation live!

This slide deck comes as close as any I have seen to teaching category theory as you would a natural language. Not too close but closer than others.

Think about it. When you entered school did the teacher begin with the terminology of grammar and how rules of grammar fit together?

Or, did the teacher start you off with “See Jack run.” or its equivalent in your language?

You were well on your way to being a competent language user before you were tasked with learning the rules for that language.

Interesting that the exact opposite approach is taken with category theory and so many topics related to computer science.

Pointers to anyone using a natural language teaching approach for category theory or CS material?

Integer sequence discovery from small graphs

Filed under: Graph Generator,Graphs,Mathematics — Patrick Durusau @ 11:36 am

Integer sequence discovery from small graphs by Travis Hoppe and Anna Petrone.

Abstract:

We have exhaustively enumerated all simple, connected graphs of a finite order and have computed a selection of invariants over this set. Integer sequences were constructed from these invariants and checked against the Online Encyclopedia of Integer Sequences (OEIS). 141 new sequences were added and 6 sequences were appended or corrected. From the graph database, we were able to programmatically suggest relationships among the invariants. It will be shown that we can readily visualize any sequence of graphs with a given criteria. The code has been released as an open-source framework for further analysis and the database was constructed to be extensible to invariants not considered in this work.

See also:

Encyclopedia of Finite Graphs “Set of tools and data to compute all known invariants for simple connected graphs”

Simple-connected-graph-invariant-database “The database file for the Encyclopedia of Finite Graphs (simple connected graphs up to order 10)”

From the paper:

A graph invariant is any property that is preserved under isomorphism. Invariants can be simple binary properties (planarity), integers (automorphism group size), polynomials (chromatic polynomials), rationals (fractional chromatic numbers), complex numbers (adjacency spectra), sets (dominion sets) or even graphs themselves (subgraph and minor matching).

Hmmm, perhaps an illustration, also from the paper, might explain better:

invariant

Figure 1: An example query to the Encyclopedia using speci c invariant conditions. The command python viewer.py 10 -i is bipartite 1 -i is integral 1 -i is eulerian 1 displays the three graphs that are simultaneously bipartite, integral and Eulerian with ten vertices.

For analysis of “cells” of your favorite evil doers you can draw higgly-piggly graphs with nodes and arcs on a whiteboard or you can take advantage of formal analysis of small graphs, research on the organization of small groups, and the history of that particular group. With the higgly-piggly approach you risk missing connections that aren’t possible to represent from your starting point.

January 27, 2015

Eigenvectors and eigenvalues: Explained Visually

Filed under: Mathematics,Visualization — Patrick Durusau @ 2:42 pm

Eigenvectors and eigenvalues: Explained Visually by Victor Powell and Lewis Lehe

Very impressive explanation/visualization of eigenvectors and eigenvalues. What is more, it concludes with pointers to additional resources.

This is only a part of a larger visualization of algorithms projects at: Explained Visually.

Looking forward to seeing more visualizations on this site.

January 6, 2015

Math, Choices, Power and Fractals

Filed under: Fractals,Mathematical Reasoning,Mathematics — Patrick Durusau @ 4:49 pm

How to Fold a Julia Fractal by Steven Wittens.

From the post:

Mathematics has a dirty little secret. Okay, so maybe it’s not so dirty. But neither is it little. It goes as follows:

Everything in mathematics is a choice.

You’d think otherwise, going through the modern day mathematics curriculum. Each theorem and proof is provided, each formula bundled with convenient exercises to apply it to. A long ladder of subjects is set out before you, and you’re told to climb, climb, climb, with the promise of a payoff at the end. “You’ll need this stuff in real life!”, they say, oblivious to the enormity of this lie, to the fact that most of the educated population walks around with “vague memories of math class and clear memories of hating it.”

Rarely is it made obvious that all of these things are entirely optional—that mathematics is the art of making choices so you can discover what the consequences are. That algebra, calculus, geometry are just words we invented to group the most interesting choices together, to identify the most useful tools that came out of them. The act of mathematics is to play around, to put together ideas and see whether they go well together. Unfortunately that exploration is mostly absent from math class and we are fed pre-packaged, pre-digested math pulp instead.

Even if you are not interested in fractals or mathematics, this is a must read post! The graphics and design of the page have to be seen to be believed. Deeply impressive!

If you are interested in fractals or mathematics, you will be stunned by the presentation in this post.

I am going to study the techniques used on this page. I don’t know if they will work with WordPress but if they don’t, I will create HTML pages and link to them from here. Not all the time but for subjects that would benefit from it.

Steven Strogatz says in the tweet I followed to this page:

This is SO good: acko.net/blog/how-to-fo… You’ll understand imaginary & complex #’s, waves, Julia and Mandelbrot sets as never before

That is so true.

BTW, did you catch the “secret” of mathematics above?

Everything in mathematics is a choice.

Very important to remember when anyone says: “The data says/shows/proves….”

BS. The data doesn’t say anything. The data plus your choices in mathematics gives X result. Not quite the same thing as “The data says/shows/proves….”

In a data driven world, only the powerless will be unable to challenge both data and the choices applied to it.

Which do you want to be? Powerful or powerless?

January 4, 2015

Linear Algebra for Machine Learning

Filed under: Machine Learning,Mathematics — Patrick Durusau @ 4:55 pm

Linear Algebra for Machine Learning by Jason Brownlee.

From the post:

You do not need to learn linear algebra before you get started in machine learning, but at some time you may wish to dive deeper.

In fact, if there was one area of mathematics I would suggest improving before the others, it would be linear algebra. It will give you the tools to help you with the other areas of mathematics required to understand and build better intuitions for machine learning algorithms.

In this post we take a closer look at linear algebra and why you should make the time to improve your skills and knowledge in linear algebra if you want to get more out of machine learning.

If you already know your way around Eigen Vectors and SVD decompositions, this post is probably not for you.

Another great collection of resources from Jason!

As usual, a great collection of resources is only the starting point for learning. The next step requires effort from the user. Sorry, wish I had better news. 😉

On the upside though, rather than thinking of it as boring mathematics, imagine how you can manipulate machine learning if you know linear algebra.

Embedding linear algebra in a machine learning book that is written from a battle perspective between different camps could be quite engaging. For that matter if online, exercises could be part of an e-warfare environment.

Something to think about.

December 11, 2014

Book of Proof

Filed under: Mathematical Reasoning,Mathematics — Patrick Durusau @ 6:52 am

Book of Proof by Richard Hammack.

From the webpage:

This book is an introduction to the standard methods of proving mathematical theorems. It has been approved by the American Institute of Mathematics’ Open Textbook Initiative. Also see the Mathematical Association of America Math DL review (of the 1st edition), and the Amazon reviews.

The second edition is identical to the first edition, except some mistakes have been corrected, new exercises have been added, and Chapter 13 has been extended. (The Cantor-Bernstein-Schröeder theorem has been added.) The two editions can be used interchangeably, except for the last few pages of Chapter 13. (But you can download them here.)

Order a copy from Amazon or Barnes & Noble for $13.75 or download a pdf for free here. Click here for a pdf copy of the entire book, or get the chapters individually below.

From the Introduction:

This book will initiate you into an esoteric world. You will learn and apply the methods of thought that mathematicians use to verify theorems, explore mathematical truth and create new mathematical theories. This will prepare you for advanced mathematics courses, for you will be better able to understand proofs, write your own proofs and think critically and inquisitively about mathematics.

For a 300+ page book, almost a steal at Amazon for $13.75. A stocking stuffer for Math/CS types on your holiday list. For yourself, grab the pdf version. 😉

Big data projects are raising the bar for being able to think critically about data and the mathematics that underlie its processing.

Big data is by definition too large for human inspection. So you had better be able to think critically about the nature of the data en masse and the methods to be used to process it.

Or to put it another way, if you don’t understand the impact of the data on processing, or assumptions built into the processing methods, how are you going to evaluate big data results?

Just accept them as ground level truth? Ignore them if they contradict your “gut?” Use a Magic 8-Ball?, a Ouija Board?

I would recommend none of the above and working on your data and math critical evaluation skills.

You?

I first saw this in a tweet by David Higginbotham.

November 10, 2014

SVM – Understanding the math

Filed under: Machine Learning,Mathematics,Support Vector Machines — Patrick Durusau @ 4:13 pm

SVM – Understanding the math – Part 1 by Alexandre Kowalczy. (Part 2)

The first two tutorials of a series on Support Vector Machines (SVM) and their use in data analysis.

If you shudder when you read:

The objective of a support vector machine is to find the optimal separating hyperplane which maximizes the margin of the training data.

you won’t after reading these tutorials. Well written and illustrated.

If you think about it, math symbolism is like programming. It is a very precise language written with a great deal of economy. Which makes it hard to understand for the uninitiated. The underlying ideas, however, can be extracted and explained. That is what you find here.

Want to improve your understanding of what appears on the drop down menu as SVM? This is a great place to start!

PS: A third tutorial is due out soon

November 6, 2014

Extracting insights from the shape of complex data using topology

Filed under: Data Analysis,Mathematics,Topological Data Analysis,Topology — Patrick Durusau @ 7:29 pm

Extracting insights from the shape of complex data using topology by P. Y. Lum, et al. (Scientific Reports 3, Article number: 1236 doi:10.1038/srep01236)

Abstract:

This paper applies topological methods to study complex high dimensional data sets by extracting shapes (patterns) and obtaining insights about them. Our method combines the best features of existing standard methodologies such as principal component and cluster analyses to provide a geometric representation of complex data sets. Through this hybrid method, we often find subgroups in data sets that traditional methodologies fail to find. Our method also permits the analysis of individual data sets as well as the analysis of relationships between related data sets. We illustrate the use of our method by applying it to three very different kinds of data, namely gene expression from breast tumors, voting data from the United States House of Representatives and player performance data from the NBA, in each case finding stratifications of the data which are more refined than those produced by standard methods.

In order to identify subjects you must first discover them.

Does the available financial contribution data on members of the United States House of Representatives correspond with the clustering analysis here? (Asking because I don’t know but would be interested in finding out.)

I first saw this in a tweet by Stian Danenbarger.

October 30, 2014

Books: Inquiry-Based Learning Guides

Filed under: Mathematics,Teaching — Patrick Durusau @ 6:09 pm

Books: Inquiry-Based Learning Guides

From the webpage:

The DAoM library includes 11 inquiry-based books freely available for classroom use. These texts can be used as semester-long content for themed courses (e.g. geometry, music and dance, the infinite, games and puzzles), or individual chapters can be used as modules to experiment with inquiry-based learning and to help supplement typical topics with classroom tested, inquiry based approaches (e.g. rules for exponents, large numbers, proof). The topic index provides an overview of all our book chapters by topic.

From the about page:

Discovering the Art of Mathematics (DAoM), is an innovative approach to teaching mathematics to liberal arts and humanities students, that offers the following vision:

Mathematics for Liberal Arts students will be actively involved in authentic mathematical experiences that

  • are both challenging and intellectually stimulating,
  • provide meaningful cognitive and metacognitive gains, and,
  • nurture healthy and informed perceptions of mathematics, mathematical ways of thinking, and the ongoing impact of mathematics not only on STEM fields but also on the liberal arts and humanities.

DAoM provides a wealth of resources for mathematics faculty to help realize this vision in their Mathematics for Liberal Arts (MLA) courses: a library of 11 inquiry-based learning guides, extensive teacher resources and many professional development opportunities. These tools enable faculty to transform their classrooms to be responsive to current research on learning (e.g. National Academy Press’s How People Learn) and the needs and interests of MLA students without enormous start-up costs or major restructuring.

All of these books are concerned with mathematics from a variety of perspectives but I didn’t see anything in How People Learn: Brain, Mind, Experience, and School: Expanded Edition (2000) that suggested such techniques are limited to the teaching of mathematics.

Easy to envision teaching of CS or semantic technologies using the same methods.

What inquiries would you construct for the exploration of semantic diversity? Roles? Contexts? Or the lack of a solution to semantic diversity? What are its costs?

Thinking semantic integration could become a higher priority if the costs of semantic diversity or the savings of semantic integration could be demonstrated.

For example, most Americans nod along with public service energy conservation messages. Just like people do with semantic integration pitches.

But if it was demonstrated for a particular home that 1/8 of the energy for heat or cooling was being wasted and that $X investment would lower utility bills by $N, there would be a much different reaction.

There are broad numbers on the losses from semantic diversity but broad numbers are not “in our budget” line items. It’s time to develop strategies that can expose the hidden costs of semantic diversity. Perhaps inquiry-based learning could be that tool.

I first saw this in a tweet by Steven Strogatz.

October 27, 2014

On the Computational Complexity of MapReduce

Filed under: Algorithms,Complexity,Computer Science,Hadoop,Mathematics — Patrick Durusau @ 6:54 pm

On the Computational Complexity of MapReduce by Jeremy Kun.

From the post:

I recently wrapped up a fun paper with my coauthors Ben Fish, Adam Lelkes, Lev Reyzin, and Gyorgy Turan in which we analyzed the computational complexity of a model of the popular MapReduce framework. Check out the preprint on the arXiv.

As usual I’ll give a less formal discussion of the research here, and because the paper is a bit more technically involved than my previous work I’ll be omitting some of the more pedantic details. Our project started after Ben Moseley gave an excellent talk at UI Chicago. He presented a theoretical model of MapReduce introduced by Howard Karloff et al. in 2010, and discussed his own results on solving graph problems in this model, such as graph connectivity. You can read Karloff’s original paper here, but we’ll outline his model below.

Basically, the vast majority of the work on MapReduce has been algorithmic. What I mean by that is researchers have been finding more and cleverer algorithms to solve problems in MapReduce. They have covered a huge amount of work, implementing machine learning algorithms, algorithms for graph problems, and many others. In Moseley’s talk, he posed a question that caught our eye:

Is there a constant-round MapReduce algorithm which determines whether a graph is connected?

After we describe the model below it’ll be clear what we mean by “solve” and what we mean by “constant-round,” but the conjecture is that this is impossible, particularly for the case of sparse graphs. We know we can solve it in a logarithmic number of rounds, but anything better is open.

In any case, we started thinking about this problem and didn’t make much progress. To the best of my knowledge it’s still wide open. But along the way we got into a whole nest of more general questions about the power of MapReduce. Specifically, Karloff proved a theorem relating MapReduce to a very particular class of circuits. What I mean is he proved a theorem that says “anything that can be solved in MapReduce with so many rounds and so much space can be solved by circuits that are yae big and yae complicated, and vice versa.

But this question is so specific! We wanted to know: is MapReduce as powerful as polynomial time, our classical notion of efficiency (does it equal P)? Can it capture all computations requiring logarithmic space (does it contain L)? MapReduce seems to be somewhere in between, but it’s exact relationship to these classes is unknown. And as we’ll see in a moment the theoretical model uses a novel communication model, and processors that never get to see the entire input. So this led us to a host of natural complexity questions:

  1. What computations are possible in a model of parallel computation where no processor has enough space to store even one thousandth of the input?
  2. What computations are possible in a model of parallel computation where processor’s can’t request or send specific information from/to other processors?
  3. How the hell do you prove that something can’t be done under constraints of this kind?
  4. How do you measure the increase of power provided by giving MapReduce additional rounds or additional time?

These questions are in the domain of complexity theory, and so it makes sense to try to apply the standard tools of complexity theory to answer them. Our paper does this, laying some brick for future efforts to study MapReduce from a complexity perspective.

Given the prevalence of MapReduce, progress on understanding what is or is not possible is an important topic.

The first two complexity questions strike me as the ones most relevant to topic map processing with map reduce. Depending upon the nature of your merging algorithm.

Enjoy!

October 10, 2014

Lance’s Lesson – Gödel Incompleteness

Filed under: Mathematical Reasoning,Mathematics,Philosophy — Patrick Durusau @ 3:45 pm

Lance’s Lesson – Gödel Incompleteness by Lance Fortnow.

The “entertainment” category on YouTube is very flexible since it included this lesson on Gödel Incompleteness. 😉

Lance uses Turing machines to “prove” the first and second incompleteness theorems in under a page of notation.

October 5, 2014

Gödel for Goldilocks…

Filed under: Mathematics,Philosophy — Patrick Durusau @ 3:27 pm

Gödel for Goldilocks: A Rigorous, Streamlined Proof of Gödel’s First Incompleteness Theorem, Requiring Minimal Background by Dan Gusfield.

Abstract:

Most discussions of Gödel’s theorems fall into one of two types: either they emphasize perceived philosophical “meanings” of the theorems, and maybe sketch some of the ideas of the proofs, usually relating Gödel’s proofs to riddles and paradoxes, but do not attempt to present rigorous, complete proofs; or they do present rigorous proofs, but in the traditional style of mathematical logic, with all of its heavy notation and difficult definitions, and technical issues which reflect Gödel’s original exposition and needed extensions by Gödel’s contemporaries. Many non-specialists are frustrated by these two extreme types of expositions and want a complete, rigorous proof that they can understand. Such an exposition is possible, because many people have realized that Gödel’s first incompleteness theorem can be rigorously proved by a simpler middle approach, avoiding philosophical discussions and hand-waiving at one extreme; and also avoiding the heavy machinery of traditional mathematical logic, and many of the harder detail’s of Gödel’s original proof, at the other extreme. This is the just-right Goldilocks approach. In this exposition we give a short, self-contained Goldilocks exposition of Gödel’s first theorem, aimed at a broad audience.

Proof that even difficult subjects can be explained without “hand=waiving” or “heavy machinery of traditional mathematical logic.”

I first saw this in a tweet by Lars Marius Garshol.

September 21, 2014

A Closed Future for Mathematics?

Filed under: Mathematica,Mathematics,Wolfram Language,WolframAlpha — Patrick Durusau @ 3:18 pm

A Closed Future for Mathematics? by Eric Raymond.

From the post:

In a blog post on Computational Knowledge and the Future of Pure Mathematics Stephen Wolfram lays out a vision that is in many ways exciting and challenging. What if all of mathematics could be expressed in a common formal notation, stored in computers so it is searchable and amenable to computer-assisted discovery and proof of new theorems?

… to be trusted, the entire system will need to be transparent top to bottom. The design, the data representations, and the implementation code for its software must all be freely auditable by third-party mathematical topic experts and mathematically literate software engineers.

Eric identifies three (3) types of errors that may exist inside the proposed closed system from Wolfram.

Is transparency of a Wolfram solution the only way to trust a Wolfram solution?

For any operation or series of operations performed with Wolfram software, you could perform the same operation in one or more open or closed source systems and see if the results agree. The more often they agree for some set of operations the greater your confidence in those operations with Wolfram software.

That doesn’t mean that the next operation or a change in the order of operations is going to produce a trustworthy result. Just that for some specified set of operations in a particular order with specified data that you obtained the same result from multiple software solutions.

It could be that all the software solutions implement the same incorrect algorithm, the same valid algorithm incorrectly, or errors in search engines searching a mathematical database (which could only be evaluated against the data being searched).

Where N is the number of non-Wolfram software packages you are using to check the Wolfram-based solution and W represents the amount of work to obtain a solution, the total work required is N x W.

In addition to not resulting in the trust Eric is describing, it is an increase in your workload.

I first saw this in a tweet by Michael Nielsen.

September 17, 2014

Elementary Applied Topology

Filed under: Mathematics,Topology — Patrick Durusau @ 7:07 pm

Elementary Applied Topology by Robert Ghrist.

From the introduction:

What topology can do

Topology was built to distinguish qualitative features of spaces and mappings. It is good for, inter alia:

  1. Characterization: Topological properties encapsulate qualitative signatures. For example, the genus of a surface, or the number of connected components of an object, give global characteristics important to classification.
  2. Continuation: Topological features are robust. The number of components or holes is not something that should change with a small error in measurement. This is vital to applications in scientific disciplines, where data is never noisy.
  3. Integration: Topology is the premiere tool for converting local data into global properties. As such, it is rife with principles and tools (Mayer-Vietoris, Excision, spectral sequences, sheaves) for integrating from local to global.
  4. Obstruction: Topology often provides tools for answering feasibility of certain problems, even when the answers to the problems themselves are hard to compute. These characteristics, classes, degrees, indices, or obstructions take the form of algebraic-topological entities.

What topology cannot do

Topology is fickle. There is no resource to tweaking epsilons should desiderata fail to be found. If the reader is a scientist or applied mathematician hoping that topological tools are a quick fix, take this text with caution. The reward of reading this book with care may be limited to the realization of new questions as opposed to new answers. It is not uncommon that a new mathematical tool contributes to applications not by answering a pressing question-of-the-day but by revealing a different (and perhaps more significant) underlying principle.

The text will require more than casual interest but what a tool to add to your toolbox!

I first saw this in a tweet from Topology Fact.

September 16, 2014

Stephen Wolfram Launching Today: Mathematica Online! (w/ secret pricing)

Filed under: Mathematica,Mathematics — Patrick Durusau @ 6:52 pm

Launching Today: Mathematica Online! by Stephen Wolfram.

From the post:

It’s been many years in the making, and today I’m excited to announce the launch of Mathematica Online: a version of Mathematica that operates completely in the cloud—and is accessible just through any modern web browser.

In the past, using Mathematica has always involved first installing software on your computer. But as of today that’s no longer true. Instead, all you have to do is point a web browser at Mathematica Online, then log in, and immediately you can start to use Mathematica—with zero configuration.

Some of the advantages that Stephen outlines:

  • Manipulate can be embedded in any web page
  • Files are stored in the Cloud to be accessed from anywhere or easily shared
  • Mathematica can now be used on mobile devices

What’s the one thing that isn’t obvious from Stephen’s post?

The pricing for access to Mathematical Online.

A Wolfram insider, proofing Stephen’s post probably said: “Oh, shit! Our pricing information is secret! What do you say in the post?

So Stephen writes:

But get Mathematica Online too (which is easy to do—through Premier Service Plus for individuals, or a site license add-on).

You do that, or at least try to do that. If you manage to hunt down Premier Service, you will find you need an activation key in order to possibly get the pricing information.

If you don’t have a copy of Mathematica, you aren’t going to be ordering Mathematica Online today.

Sad that such remarkable software has such poor marketing.

Shout out to Stephen: Lots of people are interested in using Mathematica Online or off. Byzantine marketing excludes waiting, would be paying, customers.

I first saw this in a tweet by Alex Popescu.

September 14, 2014

Building Blocks for Theoretical Computer Science

Filed under: Computer Science,Mathematics,Programming — Patrick Durusau @ 7:06 pm

Building Blocks for Theoretical Computer Science by Margaret M. Fleck.

From the preface:

This book teaches two different sorts of things, woven together. It teaches you how to read and write mathematical proofs. It provides a survey of basic mathematical objects, notation, and techniques which will be useful in later computer science courses. These include propositional and predicate logic, sets, functions, relations, modular arithmetic, counting, graphs, and trees. And, finally, it gives a brief introduction to some key topics in theoretical computer science: algorithm analysis and complexity, automata theory, and computability.

To whet your interest:

Enjoy!

I first saw this in Nat Torkington’s Four short links: 11 September 2014.

September 10, 2014

Where Does Scope Come From?

Filed under: Computer Science,Mathematics — Patrick Durusau @ 4:29 pm

Where Does Scope Come From? by Michael Robert Bernstein.

From the post:

After several false starts, I finally sat down and watched the first of Frank Pfenning’s 2012 “Proof theory foundations” talks from the University of Oregon Programming Languages Summer School (OPLSS). I am very glad that I did.

Pfenning starts the talk out by pointing out that he will be covering the “philosophy” branch of the “holy trinity” of Philosophy, Computer Science and Mathematics. If you want to “construct a logic,” or understand how various logics work, I can’t recommend this video enough. Pfenning demonstrates the mechanics of many notions that programmers are familiar with, including “connectives” (conjunction, disjunction, negation, etc.) and scope.

Scope is demonstrated during this process as well. It turns out that in logic, as in programming, the difference between a sensible concept of scope and a tricky one can often mean the difference between a proof that makes no sense, and one that you can rest other proofs on. I am very interested in this kind of fundamental kernel – how the smallest and simplest ideas are absolutely necessary for a sound foundation in any kind of logical system. Scope is one of the first intuitions that new programmers build – can we exploit this fact to make the connections between logic, math, and programming clearer to beginners? (emphasis in the original)

Michael promises more detail on the treatment of scope in future posts.

The lectures run four (4) hours so it is going to take a while to do all of them. My curiosity is whether “scope” in this context refers to variables in programming or does “scope” here extend in some way to scope as used in topic maps?

More to follow.

September 7, 2014

Matrix Methods & Applications (DRAFT)

Filed under: Mathematics,Matrix — Patrick Durusau @ 3:55 pm

Matrix Methods & Applications (DRAFT)

Stephen Boyd (Stanford) and Lieven Vandenberghe advise:

The textbook is still under very active development by Lieven Vandenberghe and Stephen Boyd, so be sure to download the newest version often. For now, we’ve posted a rough draft that does not include the exercises (which we’ll be adding). The first few chapters are in reasonable shape, but later ones are quite incomplete.

The 10 August 2014 draft has one hundred and twenty-two (122) pages so you can assume more material is coming.

I particularly like the “practical” suggested use cases.

The use cases create opportunities to illustrate the impact of data on supposedly “neutral” algorithms. Deeper knowledge of these algorithms will alert you to potential gaming of data that lies behind “neutral” processing of data.

Inspection of data is the equivalent of Mannie’s grandfather’s second rule: “Always cut cards.” (The Moon Is A Harsh Mistress)

Anyone who objects to inspection of data is hiding something. It may be their own incompetence with the data but you won’t know unless you inspect the data.

Results + algorithms + code + data = Maybe we will agree after inspection.

I first saw this in a tweet by fastml extra.

August 27, 2014

Exploring Calculus with Julia

Filed under: Julia,Mathematics,Teaching — Patrick Durusau @ 7:49 pm

Exploring Calculus with Julia

From the post:

This is a collection of notes for exploring calculus concepts with the Julia programming language. Such an approach is used in MTH 229 at the College of Staten Island.

These notes are broken into different sections, where most all sections have some self-grading questions at the end that allow you to test your knowledge of that material. The code should be copy-and-pasteable into a julia session. The code output is similar to what would be shown if evaluated in an IJulia cell, our recommended interface while learning julia.

The notes mostly follow topics of a standard first-semester calculus course after some background material is presented for learning julia within a mathematical framework.

Another example of pedagogical technique.

Semantic disconnects are legion and not hard to find. However, what criteria would you use to select a set to be solved using topic maps?

Or perhaps better, before mentioning topic maps, how would you solve them so that the solution works up to being a topic map?

Either digitally or even with pencil and paper?

Thinking that getting people to internalize the value-add of topic maps before investing effort into syntax, etc. could be a successful way to promote them.

« Newer PostsOlder Posts »

Powered by WordPress