Picture of scraped petri dish

Scrape below the surface of Strathprints...

The Strathprints institutional repository is a digital archive of University of Strathclyde research outputs. Explore world class Open Access research by researchers at Strathclyde, a leading technological university.

Explore

A multivariate correlation distance for vector spaces

Connor, Richard and Moss, Robert George (2012) A multivariate correlation distance for vector spaces. In: Similarity search and applications. Lecture Notes in Computer Science . Springer-Verlag, Berlin, pp. 209-225. ISBN 9783642321528

Full text not available in this repository. (Request a copy from the Strathclyde author)

Abstract

We investigate a distance metric, previously defined for the measurement of structured data, in the more general context of vector spaces. The metric has a basis in information theory and assesses the distance between two vectors in terms of their relative information content. The resulting metric gives an outcome based on the dimensional correlation, rather than magnitude, of the input vectors, in a manner similar to Cosine Distance. In this paper the metric is defined, and assessed, in comparison with Cosine Distance, for its major properties: semantics, properties for use within similarity search, and evaluation efficiency. We find that it is fairly well correlated with Cosine Distance in dense spaces, but its semantics are in some cases preferable. In a sparse space, it significantly outperforms Cosine Distance over TREC data and queries, the only large collection for which we have a human-ratified ground truth. This result is backed up by another experiment over movielens data. In dense Cartesian spaces it has better properties for use with similarity indices than either Cosine or Euclidean Distance. In its definitional form it is very expensive to evaluate for high-dimensional sparse vectors; to counter this, we show an algebraic rewrite which allows its evaluation to be performed more efficiently. Overall, when a multivariate correlation metric is required over positive vectors, SED seems to be a better choice than Cosine Distance in many circumstances.