scispace - formally typeset
Journal ArticleDOI

Information distance

Reads0
Chats0
TLDR
It is shown that the information distance is a universal cognitive similarity distance and investigated the maximal correlation of the shortest programs involved, the maximal uncorrelation of programs, and the density properties of the discrete metric spaces induced by the information distances.
Abstract
While Kolmogorov (1965) complexity is the accepted absolute measure of information content in an individual finite object, a similarly absolute notion is needed for the information distance between two individual objects, for example, two pictures. We give several natural definitions of a universal information metric, based on length of shortest programs for either ordinary computations or reversible (dissipationless) computations. It turns out that these definitions are equivalent up to an additive logarithmic term. We show that the information distance is a universal cognitive similarity distance. We investigate the maximal correlation of the shortest programs involved, the maximal uncorrelation of programs (a generalization of the Slepian-Wolf theorem of classical information theory), and the density properties of the discrete metric spaces induced by the information distances. A related distance measures the amount of nonreversibility of a computation. Using the physical theory of reversible computation, we give an appropriate (universal, antisymmetric, and transitive) measure of the thermodynamic work required to transform one object in another object by the most efficient process. Information distance between individual objects is needed in pattern recognition where one wants to express effective notions of "pattern similarity" or "cognitive similarity" between individual objects and in thermodynamics of computation where one wants to analyze the energy dissipation of a computation from a particular input to a particular output.

read more

Citations
More filters
Journal ArticleDOI

The Google Similarity Distance

TL;DR: A new theory of similarity between words and phrases based on information distance and Kolmogorov complexity is presented, which is applied to construct a method to automatically extract similarity, the Google similarity distance, of Words and phrases from the WWW using Google page counts.
Journal ArticleDOI

Clustering by compression

TL;DR: Evidence of successful application in areas as diverse as genomics, virology, languages, literature, music, handwritten digits, astronomy, and combinations of objects from completely different domains, using statistical, dictionary, and block sorting compressors is reported.
Posted Content

The similarity metric

TL;DR: A new "normalized information distance" is proposed, based on the noncomputable notion of Kolmogorov complexity, and it is demonstrated that it is a metric and called the similarity metric.
Journal ArticleDOI

The similarity metric

TL;DR: In this paper, the authors proposed a new normalized information distance based on the non-computable notion of Kolmogorov complexity, which minorizes every computable distance in the class (that is, it is universal in that it discovers all computable similarities).
Journal ArticleDOI

Simplicity: a unifying principle in cognitive science?

TL;DR: Research exploring the idea that simplicity drives a wide range of cognitive processes is reviewed, outlining mathematical theory, computational results and empirical data that underpin this viewpoint.
References
More filters
Journal ArticleDOI

Irreversibility and heat generation in the computing process

TL;DR: Two simple, but representative, models of bistable devices are subjected to a more detailed analysis of switching kinetics to yield the relationship between speed and energy dissipation, and to estimate the effects of errors induced by thermal fluctuations.
Journal ArticleDOI

Logical reversibility of computation

TL;DR: This result makes plausible the existence of thermodynamically reversible computers which could perform useful computations at useful speed while dissipating considerably less than kT of energy per logical step.
Journal ArticleDOI

The thermodynamics of computation—a review

TL;DR: In this paper, the authors consider the problem of rendering a computation logically reversible (e.g., creation and annihilation of a history file) in a Brownian computer, and show that it is not the making of a measurement that prevents the demon from breaking the second law but rather the logically irreversible act of erasing the record of one measurement to make room for the next.
Journal ArticleDOI

A Theory of Program Size Formally Identical to Information Theory

TL;DR: A new definition of program-size complexity is made, which has precisely the formal properties of the entropy concept of information theory.
Journal ArticleDOI

The complexity of finite objects and the development of the concepts of information and randomness by means of the theory of algorithms

TL;DR: The present article is a survey of the fundamental results connected with the concept of complexity as the minimum number of binary signs containing all the information about a given object that are sufficient for its recovery (decoding).