Another Word For It Patrick Durusau on Topic Maps and Semantic Diversity

June 7, 2012

Predictive Analytics: Decision Tree and Ensembles [part 5]

Filed under: Ensemble Methods,Machine Learning — Patrick Durusau @ 2:17 pm

Predictive Analytics: Decision Tree and Ensembles by Ricky Ho.

From the post:

Continue from my last post of walking down the list of machine learning technique. In this post, I will covered Decision Tree and Ensemble methods. We’ll continue using the iris data we prepare in this earlier post.

Ricky covers Decision Tree to illustrate early machine learning and continue under Ensemble methods to cover Random Forest and Gradient Boosted Trees.

Ricky’s next post will cover performance of the methods he has discussed in this series of posts.

June 3, 2012

Predictive Analytics: Generalized Linear Regression [part 3]

Filed under: Linear Regression,Machine Learning,Predictive Analytics — Patrick Durusau @ 3:41 pm

Predictive Analytics: Generalized Linear Regression by Ricky Ho.

From the post:

In the previous 2 posts, we have covered how to visualize input data to explore strong signals as well as how to prepare input data to a form that is situation for learning. In this and subsequent posts, I’ll go through various machine learning techniques to build our predictive model.

  1. Linear regression
  2. Logistic regression
  3. Linear and Logistic regression with regularization
  4. Neural network
  5. Support Vector Machine
  6. Naive Bayes
  7. Nearest Neighbor
  8. Decision Tree
  9. Random Forest
  10. Gradient Boosted Trees

There are two general types of problems that we are interested in this discussion; Classification is about predicting a category (value that is discrete, finite with no ordering implied) while Regression is about predicting a numeric quantity (value is continuous, infinite with ordering).

For classification problem, we use the “iris” data set and predict its “species” from its “width” and “length” measures of sepals and petals. Here is how we setup our training and testing data.

Ricky walks you through linear regression, logistic regression and linear and logistic regression with regularization.

June 2, 2012

High-Performance Domain-Specific Languages using Delite

Filed under: Delite,DSL,Machine Learning,Parallel Programming,Scala — Patrick Durusau @ 12:50 pm

High-Performance Domain-Specific Languages using Delite

Description:

This tutorial is an introduction to developing domain specific languages (DSLs) for productivity and performance using Delite. Delite is a Scala infrastructure that simplifies the process of implementing DSLs for parallel computation. The goal of this tutorial is to equip attendees with the knowledge and tools to develop DSLs that can dramatically improve the experience of using high performance computation in important scientific and engineering domains. In the first half of the day we will focus on example DSLs that provide both high-productivity and performance. In the second half of the day we will focus on understanding the infrastructure for implementing DSLs in Scala and developing techniques for defining good DSLs.

The graph manipulation language Green-Marl is one of the subjects of this tutorial.

This resource should be located and “boosted” by a search engine tuned to my preferences.

Skipping breaks, etc., you will find:

  • Introduction To High Performance DSLs (Kunle Olukotun)
  • OptiML: A DSL for Machine Learning (Arvind Sujeeth)
  • Liszt: A DSL for solving mesh-based PDEs (Zach Devito)
  • Green-Marl: A DSL for efficient Graph Analysis (Sungpack Hong)
  • Scala Tutorial (Hassan Chafi)
  • Delite DSL Infrastructure Overview (Kevin Brown)
  • High Performance DSL Implementation Using Delite (Arvind Sujeeth)
  • Future Directions in DSL Research (Hassan Chafi)

Compare your desktop computer to the MANIAC 1 (calculations for the first hydrogen bomb).

What have you invented/discovered lately?

Fuzzy machine learning framework v1.2

Filed under: Fuzzy Logic,Machine Learning — Patrick Durusau @ 9:48 am

Fuzzy machine learning framework v1.2

From the announcement:

The software is a library as well as a GTK GUI front-end for machine learning projects. Features:

  • Based on intuitionistic fuzzy sets and the possibility theory;
  • Features are fuzzy;
  • Fuzzy classes, which may intersect and can be treated as features;
  • Numeric, enumeration features and ones based on linguistic variables;
  • Derived and evaluated features;
  • Classifiers as features for building hierarchical systems;
  • User-defined features;
  • An automatic classification refinement in case of dependent features;
  • Incremental learning;
  • Object-oriented software design;
  • Features, training sets and classifiers are extensible objects;
  • Automatic garbage collection;
  • Generic data base support (through ODBC);
  • Text I/O and HTML routines for features, training sets and classifiers;
  • GTK+ widgets for features, training sets and classifiers;
  • Examples of use.

This release is packaged for Windows, Fedora (yum) and Debian (apt). The software is public domain (licensed under GM GPL).

http://www.dmitry-kazakov.de/ada/fuzzy_ml.htm

Unless you have time to waste, I would skip the religious discussion about licensing options.

For IP issues, hire lawyers, not programmers.

May 28, 2012

Streaming Analytics: with sparse distributed representations

Streaming Analytics: with sparse distributed representations by Jeff Hawkins.

Abstract:

Sparse distributed representations appear to be the means by which brains encode information. They have several advantageous properties including the ability to encode semantic meaning. We have created a distributed memory system for learning sequences of sparse distribute representations. In addition we have created a means of encoding structured and unstructured data into sparse distributed representations. The resulting memory system learns in an on-line fashion making it suitable for high velocity data streams. We are currently applying it to commercially valuable data streams for prediction, classification, and anomaly detection In this talk I will describe this distributed memory system and illustrate how it can be used to build models and make predictions from data streams.

Slides: http://www.numenta.com/htm-overview/05-08-2012-Berkeley.pdf

Looking forward to learning more about “sparse distributed representation (SDR).”

Not certain about Jeff’s claim that matching across SDRs = semantic similarity.

Design of the SDR determines the meaning of each bit and consequently of matching.

Which feeds back into the encoders that produce the SDRs.

Other resources:

The core paper: Hierarchical Temporal Memory including HTM Cortical Learning Algorithms. Check the FAQ link if you need the paper in Chinese, Japanese, Korean, Portuguese, Russian, or Spanish. (unverified translations)

Grok – Frequently Asked Questions

A very good FAQ that goes a long way to explaining the capabilities and limitations (currently) of Grok. “Unstructured text” for example isn’t appropriate input into Grok.

Jeff Hawkins and Sandra Blakeslee co-authored On Intelligence in 2004. The FAQ describes the current work as an extension of “On Intelligence.”

BTW, if you think you have heard the name Jeff Hawkins before, you have. Inventor of the Palm Pilot among other things.

May 18, 2012

From Data to Knowledge: Machine-Learning with Real-time and Streaming Applications

Filed under: Data,Knowledge,Machine Learning,Stream Analytics — Patrick Durusau @ 3:06 pm

From Data to Knowledge: Machine-Learning with Real-time and Streaming Applications

From the post:

Here is the first series of videos from the UC Berkeley Conference: From Data to Knowledge: Machine-Learning with Real-time and Streaming Applications (May 7-11, 2012). Congratulations to the local Organizing committee: Joshua Bloom, Damian Eads, Berian James, Peter Nugent, John Rice, Joseph Richards and Dan Starr for making the meeting happen and putting it all on videos for others to learn from. (in near real time!).The titles of the talks are linked to the presentation slides. The full program which ends tomorrow is here.. All the videos are here: Part 1, Part 2, Part 3, Part 4, Part 5.

Posted by Igor Carron at Nuit Blanche.

Finding enough hours to watch all of these is going to be a problem!

Which ones do you like best?

May 17, 2012

Exploring The Universe with Machine Learning

Filed under: Astroinformatics,BigData,Machine Learning — Patrick Durusau @ 3:49 pm

Exploring The Universe with Machine Learning

Webinar: Wednesday, May 30, 2012 9:00 AM – 10:00 AM (Pacific Daylight Time), (4:00pm GMT)

From the post:

WHAT IT’S ABOUT:

There is much to discover in the big, actually astronomically big, datasets that are (and will be) available. The challenge is how to effectively mine these massive datasets.

In this webinar attendees will learn how CANFAR (the Canadian Advanced Network for Astronomical Research) is using Skytree’s high performance and scalable machine learning system in the cloud. The combination enables astronomers to focus on their analyses rather than having to waste time implementing scalable complex algorithms and architecting the infrastructure to handle the massive datasets involved.

CANFAR is designed with usability in mind. Implemented as a virtual machine (VM), users can deploy their existing desktop code to the CANFAR cloud – delivering instant scalability (replication of the VM as required), without additional development.

WHO SHOULD ATTEND:

Anyone interested in performing machine learning or advanced analytics on big (astronomical) data sets.

Well, I quality on two counts. How about you? 😉

From Skytree Big Data Analytics. They have a free server version that I haven’t looked at, yet.

May 5, 2012

DARPA system to blend AI, machine learning to understand mountain of text

Filed under: Artificial Intelligence,Machine Learning — Patrick Durusau @ 6:55 pm

DARPA system to blend AI, machine learning to understand mountain of text

From the post:

The Defense Advanced Research Projects Agency (DARPA) will next this month detail the union of advanced technologies from artificial intelligence, computational linguistics, machine learning, natural-language fields it hopes to bring together to build an automated system that will let analysts and others better grasp meanings from large volumes of text documents.

From DARPA: “Automated, deep natural-language understanding technology may hold a solution for more efficiently processing text information. When processed at its most basic level without ingrained cultural filters, language offers the key to understanding connections in text that might not be readily apparent to humans. Sophisticated artificial intelligence of this nature has the potential to enable defense analysts to efficiently investigate orders of magnitude more documents so they can discover implicitly expressed, actionable information contained within them.”

DARPA is holding a proposers day, May 16, 2012 in Arlington, VA, on the Deep Exploration and Filtering of Text (DEFT) project.

I won’t be attending but am interested in what you learn about the project.

What has me curious is that assuming DEFT is successful, how do they intend to capture the insights of analysts who describe the data and their conclusions differently? Particularly over time or from the perspective of different intelligence agencies? Or document the trails a particular analyst has followed through a mountain of data? Seems like those would be important issues as well.

Issues that are uniquely suited for subject-centric approaches like topic maps.

May 4, 2012

Machine Learning in Python Has Never Been Easier!

Filed under: Machine Learning,Python — Patrick Durusau @ 3:41 pm

Machine Learning in Python Has Never Been Easier!

From the post:

At BigML we believe that over the next few years automated, data-driven decisions and data-driven applications are going to change the world. In fact, we think it will be the biggest shift in business efficiency since the dawn of the office calculator, when individuals had “Computer” listed as the title on their business card. We want to help people rapidly and easily create predictive models using their datasets, no matter what size they are. Our easy-to-use, public API is a great step in that direction but a few bindings for popular languages is obviously a big bonus.

Thus, we are very happy to announce an open source Python binding to BigML.io, the BigML REST API. You can find it and fork it at Github.

The BigML Python module makes it extremely easy to programmatically manage BigML sources, datasets, models and predictions. The snippet below sketches how you can create a source, dataset, model and then a prediction for a new object.

The “business efficiency” argument sounds like the “paperless office” to me.

Certain we will be able to do different, interesting and quite possibly useful things with machine learning and data. That we will become more “efficient,” is a separate question. By what measure?

If you look at scholarship from the 19th century, where people lacked many of the time saving devices of today, you will find authors who published hundreds of books, not articles, books. And not short books either. Were they more “efficient” than we are?

Rather than promise “efficiency,” promote machine learning as a means to do a particular task and do it well. If there is interest in the task and/or the result, that will be sufficient without all the superlatives.

Machine See, Machine Do

Filed under: Games,Machine Learning,Multimedia,Music,Music Retrieval — Patrick Durusau @ 3:40 pm

While we wait for maid service robots, news that computers can be trained as human mimics for labeling of multimedia resources. Game-powered machine learning reports success with game based training for music labeling.

The authors, Luke Barrington, Douglas Turnbull, and Gert Lanckriet, neatly summarize music labeling as a problem of volume:

…Pandora, a popular Internet radio service, employs musicologists to annotate songs with a fixed vocabulary of about five hundred tags. Pandora then creates personalized music playlists by finding songs that share a large number of tags with a user-specified seed song. After 10 y of effort by up to 50 full time musicologists, less than 1 million songs have been manually annotated (5), representing less than 5% of the current iTunes catalog.

A problem that extends to the “…7 billion images are uploaded to Facebook each month (1), YouTube users upload 24 h of video content per minute….”

The authors created www.HerdIt.org to:

… investigate and answer two important questions. First, we demonstrate that the collective wisdom of Herd It’s crowd of nonexperts can train machine learning algorithms as well as expert annotations by paid musicologists. In addition, our approach offers distinct advantages over training based on static expert annotations: it is cost-effective, scalable, and has the flexibility to model demographic and temporal changes in the semantics of music. Second, we show that integrating Herd It in an active learning loop trains accurate tag models more effectively; i.e., with less human effort, compared to a passive approach.

The approach promises an augmentation (not replacement) of human judgement with regard to classification of music. An augmentation that would enable human judgement to reach further across the musical corpus than ever before:

…while a human-only approach requires the same labeling effort for the first song as for the millionth, our game-powered machine learning solution needs only a small, reliable training set before all future examples can be labeled automatically, improving efficiency and cost by orders of magnitude. Tagging a new song takes 4 s on a modern CPU: in just a week, eight parallel processors could tag 1 million songs or annotate Pandora’s complete song collection, which required a decade of effort from dozens of trained musicologists.

A promising technique for IR with regard to multimedia resources.

What I wonder about is the extension of the technique, games designed to train machine learning for:

  • e-discovery in legal proceedings
  • “tagging” or indexing if you will, text resources
  • vocabulary expansion for searching
  • contexts for semantic matching
  • etc.

A first person shooter game that annotates the New York Times archives would be really cool!

May 3, 2012

Hyperbolic lots

Filed under: Humor,Language,Machine Learning — Patrick Durusau @ 6:23 pm

Hyperbolic lots by Ben Zimmer.

From the post:

For the past couple of years, Google has provided automatic captioning for all YouTube videos, using a speech-recognition system similar to the one that creates transcriptions for Google Voice messages. It’s certainly a boon to the deaf and hearing-impaired. But as with Google’s other ventures in natural language processing (notably Google Translate), this is imperfect technology that is gradually becoming less imperfect over time. In the meantime, however, the imperfections can be quite entertaining.

I gave the auto-captioning an admittedly unfair challenge: the multilingual trailer that Michael Erard put together for his latest book, Babel No More: The Search for the World’s Most Extraordinary Language Learners. The trailer features a story from the book told by speakers of a variety of languages (including me), and Erard originally set it up as a contest to see who could identify the most languages. If you go to the original video on YouTube, you can enable the auto-captioning by clicking on the “CC” and selecting “Transcribe Audio” from the menu.

The transcription does a decent job with Erard’s English introduction, though I enjoyed the interpretation of “hyperpolyglots” — the subject of the book — as “hyperbolic lots.” Hyperpolyglot (evidently coined by Dick Hudson) isn’t a word you’ll find in any dictionary, and it’s not that frequent online, so it’s highly unlikely the speech-to-text system could have figured it out. But the real fun begins with the speakers of other languages.

You will find this amusing.

Ben notes the imperfections are becoming fewer.

Curious, since languages are living, social constructs, at what point to we measure the number of “imperfections?”

Or should I say from whose perspective do we measure the number of “imperfections?”

Or should we use both of those measures and others?

April 25, 2012

Online tool can detect patterns in US election news coverage

Filed under: Machine Learning,News,Politics — Patrick Durusau @ 6:27 pm

Online tool can detect patterns in US election news coverage

From the website:

The US presidential election dominates the global media every four years, with news articles, which are carefully analysed by commentators and campaign strategists, playing a major role in shaping voter opinion.

Academics at the University of Bristol’s Intelligent Systems Laboratory have developed an online tool, Election Watch, which analyses the content of news about the US election by the international media.

A paper about the project will be presented at the Proceedings of the 13th conference of the European Chapter of the Association for Computational Linguistics held in Avignon, France.

Election Watch automatically monitors political discourse about the 2012 US presidential election from over 700 American and international news outlets. The information displayed is based, so far, on 91,456 articles.

The web tool allows users to explore news stories via an interactive interface and demonstrates the application of modern machine learning and language technologies. After analysing news articles about the 2012 US election the researchers have found patterns in the political narrative.

The online site is updated daily, by presenting narrative patterns as they were extracted from news. Narrative patterns include actors, actions, triplets representing political support between actors, and automatically inferred political allegiance of actors.

The site also presents the key named entities, timelines and heat maps. Network analysis allows the researchers to infer the role of each actor in the general political discourse, recognising adversaries and allied actors. Users can browse articles by political statements, rather than by keywords. For example, users can browse articles where Romney is described as criticising Obama. All the graphical briefing is automatically generated and interactive and each relation presented to the user can be used to retrieve supporting articles, from a set of hundreds of online news sources.

You really have to see this website. Quite amazing.

I would disagree with the placement of Obama to the far left in at least one of the graphics.

From where I sit he should be cheek and jowl with Romney, albeit on his left side.

I wonder if the data set is going to be released or if that is possible?

PBS should ask permission to carry this in a frame on their site.

LAILAPS

LAILAPS

From the website:

LAILAPS combines a keyword driven search engine for an integrative access to life science databases, machine learning for a content driven relevance ranking, recommender systems for suggestion of related data records and query refinements with a user feedback tracking system for an self learning relevance training.

Features:

  • ultra fast keyword based search
  • non-static relevance ranking
  • user specific relevance profiles
  • suggestion of related entries
  • suggestion of related query terms
  • self learning by user tracking
  • deployable at standard desktop PC
  • 100% JAVA
  • installer for in-house deployment

I like the idea of a recommender system that “suggests” related data records and query refinements. It could be wrong.

I am as guilty as anyone of thinking in terms of “correct” recommendations that always lead to relevant data.

That is applying “crisp” set thinking to what is obviously a “rough” set situation. We as readers have to sort out the items in the “rough” set and construct for ourselves, a temporary and fleeting “crisp” set for some particular purpose.

If you are using LAILAPS, I would appreciate a note about your experiences and impressions.

April 24, 2012

Machine learning for identification of cars

Filed under: Machine Learning,R — Patrick Durusau @ 7:14 pm

Machine learning for identification of cars by Dzidorius Martinaitis.

A very awesome post (with code) on capturing video from traffic cameras and training your computer to recognize cars.

The post covers using video from public sources but the thought does occur to me that you could spool the output from a digital camera attached to a telephoto lens to a personal computer for encryption and transfer over the Net. So that you could get higher quality images than off a public feed.

I am sure you will enjoy experimenting with it both as illustrated in the post and as other possibilities suggest themselves to you.

Software Review- BigML.com – Machine Learning meets the Cloud

Filed under: Cloud Computing,Machine Learning — Patrick Durusau @ 7:13 pm

Software Review- BigML.com – Machine Learning meets the Cloud.

Ajay Ohri reviews BigML.com, an attempt to lower the learning curve for working with machine learning and large data sets.

Ajay concludes:

Overall a welcome addition to make software in the real of cloud computing and statistical computation/business analytics both easy to use and easy to deploy with fail safe mechanisms built in.

Check out https://bigml.com/ for yourself to see.

I have to agree they are off to a good start.

Lowering the learning curve applications look like “hot” properties for the coming future. Some lose of flexibility but offset by immediate and possibly useful results. Maybe the push some users need to become real experts.

April 23, 2012

scikit-learn – Machine Learning in Python – Astronomy

Filed under: Astroinformatics,Machine Learning,Python — Patrick Durusau @ 5:57 pm

scikit-learn – Machine Learning in Python – Astronomy by Jake VanderPlas. (tutorial)

Jake branched the scikit-learn site for his tutorial on scikit-learn using astronomical data.

Good introduction to scikit-learn and will be of interest to astronomy buffs.

ICDM 2012

ICDM 2012 Brussels, Belgium | December 10 – 13, 2012

From the webpage:

The IEEE International Conference on Data Mining series (ICDM) has established itself as the world’s premier research conference in data mining. It provides an international forum for presentation of original research results, as well as exchange and dissemination of innovative, practical development experiences. The conference covers all aspects of data mining, including algorithms, software and systems, and applications.

ICDM draws researchers and application developers from a wide range of data mining related areas such as statistics, machine learning, pattern recognition, databases and data warehousing, data visualization, knowledge-based systems, and high performance computing. By promoting novel, high quality research findings, and innovative solutions to challenging data mining problems, the conference seeks to continuously advance the state-of-the-art in data mining. Besides the technical program, the conference features workshops, tutorials, panels and, since 2007, the ICDM data mining contest.

Important Dates:

ICDM contest proposals: April 30
Conference full paper submissions: June 18
Demo and tutorial proposals: August 10
Workshop paper submissions: August 10
PhD Forum paper submissions: August 10
Conference paper, tutorial, demo notifications: September 18
Workshop paper notifications: October 1
PhD Forum paper notifications: October 1
Camera-ready copies and copyright forms: October 15

April 22, 2012

AI & Statistics 2012

Filed under: Artificial Intelligence,Machine Learning,Statistical Learning,Statistics — Patrick Durusau @ 7:08 pm

AI & Statistics 2012 (La Palma, Canary Islands)

Proceedings:

http://jmlr.csail.mit.edu/proceedings/papers/v22/

As one big file:

http://jmlr.csail.mit.edu/proceedings/papers/v22/v22.tar.gz

Why you should care:

The fifteenth international conference on Artificial Intelligence and Statistics (AISTATS 2012) will be held on La Palma in the Canary Islands. AISTATS is an interdisciplinary gathering of researchers at the intersection of computer science, artificial intelligence, machine learning, statistics, and related areas. Since its inception in 1985, the primary goal of AISTATS has been to broaden research in these fields by promoting the exchange of ideas among them. We encourage the submission of all papers which are in keeping with this objective.

The conference runs April 21 – 23, 2012. Sorry!

You will enjoy looking over the papers!

GraphLab: Workshop on Big Learning

Filed under: Algorithms,BigData,Machine Learning — Patrick Durusau @ 7:06 pm

GraphLab: Workshop on Big Learning

Monday, July 9, 2012

From the webpage:

The GraphLab workshop on large scale machine learning is a meeting place for both academia and industry to discuss upcoming challenges of large scale machine learning and solution methods. GraphLab is Carnegie Mellon’s large scale machine learning framework. The workshop will include demos and tutorials showcasing the next generation of the GraphLab framework, as well as lectures and demos from the top technology companies about their applied large scale machine learning solutions.

and

There is a related workshop on Algorithms for Modern Massive Datasets at Stanford, immediately after the GraphLab workshop.

If you are going to be in the Bay area, definitely a good way to start the week!

April 18, 2012

DUALIST: Utility for Active Learning with Instances and Semantic Terms

Filed under: Active Learning,Bayesian Models,HCIR,Machine Learning — Patrick Durusau @ 6:08 pm

DUALIST: Utility for Active Learning with Instances and Semantic Terms

From the webpage:

DUALIST is an interactive machine learning system for quickly building classifiers for text processing tasks. It does so by asking “questions” of a human “teacher” in the form of both data instances (e.g., text documents) and features (e.g., words or phrases). It uses active learning and semi-supervised learning to build text-based classifiers at interactive speed.

(video demo omitted)

The goals of this project are threefold:

  1. A practical tool to facilitate annotation/learning in text analysis projects.
  2. A framework to facilitate research in interactive and multi-modal active learning. This includes enabling actual user experiments with the GUI (as opposed to simulated experiments, which are pervasive in the literature but sometimes inconclusive for use in practice) and exploring HCI issues, as well as supporting new dual supervision algorithms which are fast enough to be interactive, accurate enough to be useful, and might make more appropriate modeling assumptions than multinomial naive Bayes (the current underlying model).
  3. A starting point for more sophisticated interactive learning scenarios that combine multiple “beyond supervised learning” strategies. See the proceedings of the recent ICML 2011 workshop on this topic.

This could be quite useful for authoring a topic map across a corpus of materials. With interactive recognition of occurrences of subjects, etc.

Sponsored in part by the folks at DARPA. Unlike Al Gore, they did build the Internet.

April 13, 2012

Operations, machine learning and premature babies

Filed under: Bioinformatics,Biomedical,Machine Learning — Patrick Durusau @ 4:40 pm

Operations, machine learning and premature babies: An astonishing connection between web ops and medical care. By Mike Loukides.

From the post:

Julie Steele and I recently had lunch with Etsy’s John Allspaw and Kellan Elliott-McCrea. I’m not sure how we got there, but we made a connection that was (to me) astonishing between web operations and medical care for premature infants.

I’ve written several times about IBM’s work in neonatal intensive care at the University of Toronto. In any neonatal intensive care unit (NICU), every baby is connected to dozens of monitors. And each monitor is streaming hundreds of readings per second into various data systems. They can generate alerts if anything goes severely out of spec, but in normal operation, they just generate a summary report for the doctor every half hour or so.

IBM discovered that by applying machine learning to the full data stream, they were able to diagnose some dangerous infections a full day before any symptoms were noticeable to a human. That’s amazing in itself, but what’s more important is what they were looking for. I expected them to be looking for telltale spikes or irregularities in the readings: perhaps not serious enough to generate an alarm on their own, but still, the sort of things you’d intuitively expect of a person about to become ill. But according to Anjul Bhambhri, IBM’s Vice President of Big Data, the telltale signal wasn’t spikes or irregularities, but the opposite. There’s a certain normal variation in heart rate, etc., throughout the day, and babies who were about to become sick didn’t exhibit the variation. Their heart rate was too normal; it didn’t change throughout the day as much as it should.

That observation strikes me as revolutionary. It’s easy to detect problems when something goes out of spec: If you have a fever, you know you’re sick. But how do you detect problems that don’t set off an alarm? How many diseases have early symptoms that are too subtle for a human to notice, and only accessible to a machine learning system that can sift through gigabytes of data?

The post goes on to discuss how our servers may exhibit behaviors that machine learning could recognize but that we can’t specify.

That may be Rumsfeld’s “unknown unknowns,” however we all laughed at the time.

There are “unknown unknown’s” and tireless machine learning may be the only way to identify them.

In topic map lingo, I would say there are subjects that we haven’t yet learned to recognize.

April 12, 2012

From Zero to Machine Learning in Less than Seven Minutes

Filed under: Machine Learning — Patrick Durusau @ 7:05 pm

From Zero to Machine Learning in Less than Seven Minutes by Charles Parker.

From the post:

Here at BigML, we do a lot of work trying to make machine learning accessible. This involves a lot of thought about everything from classification algorithms, to data visualization, to infrastructure, databases, particle physics, and security.

Okay, not particle physics. But definitely all of that other stuff.

After all that thinking, our hope is that we’ve built something that non-experts can use to build data-driven decisions into their applications and business logic. To get you started, we’ve made a series of short videos showing the key features of the site. Watch and learn. Machine learning is only seven minutes away.

An impressively done “…less than seven minutes. Watch all the videos and in particular watch for the “live pruning slider.” Worth the time you will spend on the videos.

It elides over many of the difficulties found in machine learning, but isn’t that part of being a service? That is if you tooled this by hand, there would be a lot more detail and choices at every turn.

By reducing the number of options and choices, as well as glossing over some of the explanations, this service may bring machine learning to a larger user population.

What would it look like to do something similar for topic maps?

Thoughts?

April 11, 2012

Close Counts In Horseshoes, Hand Grenades and Clustering

Filed under: Clustering,Machine Learning,R — Patrick Durusau @ 6:18 pm

Machine Learning in R: Clustering by Ricky Ho.

Ricky writes:

Clustering is a very common technique in unsupervised machine learning to discover groups of data that are “close-by” to each other. It is broadly used in customer segmentation and outlier detection.

It is based on some notion of “distance” (the inverse of similarity) between data points and use that to identify data points that are close-by to each other. In the following, we discuss some very basic algorithms to come up with clusters, and use R as examples.

Covers K-Means, Hierarchical Clustering, Fuzzy C-Means, Multi-Gaussian with Expectation-Maximization, and Density-based Cluster algorithms.

Good introduction to the basics of clustering in R.

April 6, 2012

Is Machine Learning v Domain expertise the wrong question?

Filed under: Domain Expertise,Machine Learning — Patrick Durusau @ 6:48 pm

Is Machine Learning v Domain expertise the wrong question?

James Taylor writes:

KDNuggets had an interesting poll this week in which readers expressed themselves as Skeptical of Machine Learning replacing Domain Expertise. This struck me not because I disagree but because I think it is in some ways the wrong question:

  • Any given decision is made based on a combination of information, know-how and pre-cursor decisions.
  • The know-how can be based on policy, regulation, expertise, best practices or analytic insight (such as machine learning).
  • Some decisions are heavily influenced by policy and regulation (deciding if a claim is complete and valid for instance) while others are more heavily influenced by the kind of machine learning insight common in analytics (deciding if the claim is fraudulent might be largely driven by a Neural Network that determines how “normal” the claim seems to be).
  • Some decisions are driven primarily by the results of pre-cursor or dependent decisions.
  • All require access to some set of information.

I think the stronger point, the one that James closes with, is decision management needs machine learning and domain expertise, together.

And we find our choices of approaches justified by the results, “as we see them.” What more could you ask for?

March 31, 2012

23rd International Conference on Algorithmic Learning Theory (ALT 2012)

Filed under: Conferences,Machine Learning — Patrick Durusau @ 4:10 pm

23rd International Conference on Algorithmic Learning Theory (ALT 2012)

Important Dates:

Submission Deadline: May 17, 2012

Notification: July 8, 2012

Camera ready copy: July 20, 2012

Early registration deadline: August 30, 2012

The conference: October 29 – 31, 2012

From the call for papers:

The 23rd International Conference on Algorithmic Learning Theory (ALT 2012) will be held in Lyon, France, at Université Lumière Lyon 2, on October 29-31, 2012. The conference is on the theoretical foundations of machine learning. The conference will be co-located with the 15th International Conference on Discovery Science (DS 2012)

Topics of Interest: We invite submissions that make a wide
variety of contributions to the theory of learning, including the
following:

  • Comparison of the strength of learning models and the design and
    evaluation of novel algorithms for learning problems in
    established learning-theoretic settings such as

    • statistical learning theory,
    • on-line learning,
    • inductive inference,
    • query models,
    • unsupervised, semi-supervised and active learning.
  • Analysis of the theoretical properties of existing algorithms:
    • families of algorithms could include
      • boosting,
      • kernel-based methods, SVM,
      • Bayesian networks,
      • methods for reinforcement learning or learning in
        repeated games,

      • graph- and/or manifold-based methods,
      • methods for latent-variable estimation and/or clustering,
      • MDL,
      • decision tree methods,
      • information-based methods,
    • analyses could include generalization, convergence or
      computational efficiency.
  • Definition and analysis of new learning models. Models might
    • identify and formalize classes of learning problems
      inadequately addressed by existing theory or

    • capture salient properties of important concrete applications.

.

Curious: Do you know of any research comparing the topics of interest for a conference against the terms used in presentations for the conference?

DS 2012 : The 15th International Conference on Discovery Science

Filed under: Conferences,Data Mining,Machine Learning — Patrick Durusau @ 4:09 pm

DS 2012 : The 15th International Conference on Discovery Science

Important Dates:

Important Dates for Submissions

Full paper submission: 17 th May, 2012
Author notification: 8th July, 2012
Camera-ready papers due: 20th July, 2012

Important dates for all DS 2012 attendees

Deadline for early registration: 30th August, 2012
DS 2012 conference dates: 29-31 October, 2012

From the call for papers:

DS-2012 will be collocated with ALT-2012, the 23rd International Conference on Algorithmic Learning Theory. The two conferences will be held in parallel, and will share their invited talks.

DS 2012 provides an open forum for intensive discussions and exchange of new ideas among researchers working in the area of Discovery Science. The scope of the conference includes the development and analysis of methods for automatic scientific knowledge discovery, machine learning, intelligent data analysis, theory of learning, as well as their application to knowledge discovery. Very welcome are papers that focus on dynamic and evolving data, models and structures.

We invite submissions of research papers addressing all aspects of discovery science. We particularly welcome contributions that discuss the application of data analysis, data mining and other support techniques for scientific discovery including, but not limited to, biomedical, astronomical and other physics domains.

Possible topics include, but are not limited to:

  • Logic and philosophy of scientific discovery
  • Knowledge discovery, machine learning and statistical methods
  • Ubiquitous Knowledge Discovery
  • Data Streams, Evolving Data and Models
  • Change Detection and Model Maintenance
  • Active Knowledge Discovery
  • Learning from Text and web mining
  • Information extraction from scientific literature
  • Knowledge discovery from heterogeneous, unstructured and multimedia data
  • Knowledge discovery in network and link data
  • Knowledge discovery in social networks
  • Data and knowledge visualization
  • Spatial/Temporal Data
  • Mining graphs and structured data
  • Planning to Learn
  • Knowledge Transfer
  • Computational Creativity
  • Human-machine interaction for knowledge discovery and management
  • Biomedical knowledge discovery, analysis of micro-array and gene deletion data
  • Machine Learning for High-Performance Computing, Grid
    and Cloud Computing
  • Applications of the above techniques to natural or social sciences

I looked very briefly at prior proceedings. If those are any indication, this should be a very good conference.

March 25, 2012

Book Review- Machine Learning for Hackers

Filed under: Machine Learning,R — Patrick Durusau @ 7:14 pm

Book Review- Machine Learning for Hackers by Ajay Ohri.

From the post:

This is review of the fashionably named book Machine Learning for Hackers by Drew Conway and John Myles White (O’Reilly ). The book is about hacking code in R.

The preface introduces the reader to the authors conception of what machine learning and hacking is all about. If the name of the book was machine learning for business analytsts or data miners, I am sure the content would have been unchanged though the popularity (and ambiguity) of the word hacker can often substitute for its usefulness. Indeed the many wise and learned Professors of statistics departments through out the civilized world would be mildly surprised and bemused by their day to day activities as hacking or teaching hackers. The book follows a case study and example based approach and uses the GGPLOT2 package within R programming almost to the point of ignoring any other native graphics system based in R. It can be quite useful for the aspiring reader who wishes to understand and join the booming market for skilled talent in statistical computing.

A chapter by chapter review that highlights a number of improvements that one hopes will appear in a second (2nd) edition. Mostly editorial, clarity type improvements that should be been caught in editorial review.

The complete source code for examples can be downloaded here. It is a little over 100 MB in zip format. I checked and the data files for various exercises are included. Which explains the size of the source code file.

March 23, 2012

Excellent Papers for 2011 (Google)

Filed under: HCIR,Machine Learning,Multimedia,Natural Language Processing — Patrick Durusau @ 7:23 pm

Excellent Papers for 2011 (Google)

Corinna Cortes and Alfred Spector of Google Research have collected up great papers published by Glooglers in 2011.

To be sure there are the obligatory papers on searching and natural language processing but there are also papers on audio processing, human-computer interfaces, multimedia, systems and other topics.

Many of these will be the subjects of separate posts in the future. For now, peruse at your leisure and sing out when you see one of special interest.

March 22, 2012

Text Analytics in Telecommunications – Part 3

Filed under: Machine Learning,Text Analytics — Patrick Durusau @ 7:41 pm

Text Analytics in Telecommunications – Part 3 by Themos Kalafatis.

From the post:

It is well known that FaceBook contains a multitude of information that can be potentially analyzed. A FaceBook page contains several entries (Posts, Photos, Comments, etc) which in turn generate Likes. This data can be analyzed to better understand the behavior of consumers towards a Brand, Product or Service.

Let’s look at the analysis of the three FaceBook pages of MT:S, Telenor and VIP Mobile Telcos in Serbia as an example. The question that this analysis tries to answer is whether we can identify words and phrases that frequently appear in posts that generate any kind of reaction (a “Like”, or a Comment) vs words and topics that do not tend to generate reactions . If we are able to differentiate these words then we get an idea on what consumers tend to value more : If a post is of no value to us then we will not tend to Like it and/or comment it.

To perform this analysis we need a list of several thousands of posts (their text) and also the number of Likes and Comments that each post has received. If any post has generated a Like and/or a Comment then we flag that post as having generated a reaction. The next step is to feed that information to a machine learning algorithm to identify which words have discriminative power (=which words appear more frequently in posts that are liked and/or commented and also which words do not produce any reaction.)

It would be more helpful if the “machine learning algorithm” used in this case was identified, along with the data set in question.

I suppose we will learn more after the presentation at the European Text Analytics Summit, although we would like to learn more sooner! 😉

March 18, 2012

Learning from Data

Filed under: CS Lectures,Machine Learning — Patrick Durusau @ 8:53 pm

Learning from Data

Outline:

This is an introductory course on machine learning that covers the basic theory, algorithms and applications. Machine learning (ML) uses data to recreate the system that generated the data. ML techniques are widely applied in engineering, science, finance, and commerce to build systems for which we do not have full mathematical specification (and that covers a lot of systems). The course balances theory and practice, and covers the mathematical as well as the heuristic aspects. Detailed topics are listed below.

From the webpage:

Real Caltech course, not watered-down version
Broadcast live from the lecture hall at Caltech

And so, the competition of online course offerings begins. 😉

« Newer PostsOlder Posts »

Powered by WordPress