A word at a time: computing word relatedness using temporal semantic analysis by Kira Radinsky, Technion-Israel Institute of Technology, Haifa, Israel; Eugene Agichtein, Emory University, Atlanta, GA, USA; Evgeniy Gabrilovich, Yahoo! Research, Santa Clara, CA, USA; Shaul Markovitch, Technion-Israel Institute of Technology, Haifa, Israel.
Computing the degree of semantic relatedness of words is a key functionality of many language applications such as search, clustering, and disambiguation. Previous approaches to computing semantic relatedness mostly used static language resources, while essentially ignoring their temporal aspects. We believe that a considerable amount of relatedness information can also be found in studying patterns of word usage over time. Consider, for instance, a newspaper archive spanning many years. Two words such as “war” and “peace” might rarely co-occur in the same articles, yet their patterns of use over time might be similar. In this paper, we propose a new semantic relatedness model, Temporal Semantic Analysis (TSA), which captures this temporal information. The previous state of the art method, Explicit Semantic Analysis (ESA), represented word semantics as a vector of concepts. TSA uses a more refined representation, where each concept is no longer scalar, but is instead represented as time series over a corpus of temporally-ordered documents. To the best of our knowledge, this is the first attempt to incorporate temporal evidence into models of semantic relatedness. Empirical evaluation shows that TSA provides consistent improvements over the state of the art ESA results on multiple benchmarks.
The discovery of “related” terms may lead to discovery of synonyms for a subject, associations with a subject and other grist for your topic map mill.
This is interesting work and should be considered whenever topic mapping material recorded over time. Historical government archives come to mind.