Journal ArticleDOI
On the Complexity of Finite Sequences
A. Lempel,Jacob Ziv +1 more
Reads0
Chats0
TLDR
A new approach to the problem of evaluating the complexity ("randomness") of finite sequences is presented, related to the number of steps in a self-delimiting production process by which a given sequence is presumed to be generated.Abstract:
A new approach to the problem of evaluating the complexity ("randomness") of finite sequences is presented. The proposed complexity measure is related to the number of steps in a self-delimiting production process by which a given sequence is presumed to be generated. It is further related to the number of distinct substrings and the rate of their occurrence along the sequence. The derived properties of the proposed measure are discussed and motivated in conjunction with other well-established complexity criteria.read more
Citations
More filters
Journal ArticleDOI
Analysis and evaluation of handwriting in patients with Parkinson's disease using kinematic, geometrical, and non-linear features.
Cristian David Rios-Urrego,Juan Camilo Vásquez-Correa,Juan Camilo Vásquez-Correa,Jesús Francisco Vargas-Bonilla,Elmar Nöth,Francisco Lopera,Juan Rafael Orozco-Arroyave +6 more
TL;DR: The results confirmed the negative impact of aging in the classification process when the authors considered different groups of healthy subjects and how those features are able to discriminate between Parkinson's disease patients and healthy subjects.
Journal ArticleDOI
Quad tree structures for image compression applications
TL;DR: A class of lossy algorithms that is capable of compressing image data over a wide range of rates so that quick browsing of large amounts of information as well as detailed examination of high resolution areas can be achieved by the same compression system.
Journal ArticleDOI
Linear and nonlinear analysis of airflow recordings to help in sleep apnoea-hypopnoea syndrome diagnosis
TL;DR: The result suggests that AF and RRV provide useful information to detect SAHS, and the diagnostic performance of single features and LR models is assessed and compared in terms of sensitivity, specificity, accuracy and area under the receiver-operating characteristics curve (AROC).
Book ChapterDOI
Fast and practical algorithms for computing all the runs in a string
TL;DR: A collection of fast space-efficient algorithms for computing all the runs in a string that appear in many circumstances to be superior to those previously proposed.
Journal ArticleDOI
Consciousness is supported by near-critical slow cortical electrodynamics
Daniel Toker,Ioannis Pappas,J. Lendner,Joel Frohlich,Diego M. Mateos,Suresh D. Muthukumaraswamy,Robin L. Carhart-Harris,Michelle Paff,Paul M. Vespa,Martin M. Monti,Friedrich T. Sommer,Robert T. Knight,Mark D'Esposito +12 more
TL;DR: It is shown that the electric activity of the cortex is indeed poised near the boundary between stability and chaos during conscious states and transitions away from this boundary during unconsciousness and that this transition disrupts cortical information processing.
References
More filters
Journal ArticleDOI
Three approaches to the quantitative definition of information
TL;DR: In this article, three approaches to the quantitative definition of information are presented: information-based, information-aware and information-neutral approaches to quantifying information in the context of information retrieval.
Journal ArticleDOI
The definition of random sequences
TL;DR: It is shown that the random elements as defined by Kolmogorov possess all conceivable statistical properties of randomness and can equivalently be considered as the elements which withstand a certain universal stochasticity test.
Journal ArticleDOI
Process complexity and effective random tests
TL;DR: A variant of the Kolmogorov concept of complexity which yields a common theory of finite and infinite random sequences and some concepts of effective tests which are proved to be equivalent are established.
Journal ArticleDOI
Information-Theoretic Limitations of Formal Systems
TL;DR: An attempt is made to apply information-theoretic computational complexity to meta-mathematics by measuring the difficulty of proving a given set of theorems, in terms of the number of bits of axioms that are assumed, and the size of the proofs needed to deduce the theoremic proofs.