 Algorithmically random sequence

Intuitively, an algorithmically random sequence (or random sequence) is an infinite sequence of binary digits that appears random to any algorithm. The definition applies equally well to sequences on any finite set of characters. Random sequences are key objects of study in algorithmic information theory.
As different types of algorithms are sometimes considered, ranging from algorithms with specific bounds on their running time to algorithms which may ask questions of an oracle, there are different notions of randomness. The most common of these is known as MartinLöf randomness (or 1randomness), but stronger and weaker forms of randomness also exist. The term "random" used to refer to a sequence without clarification is usually taken to mean "MartinLöf random" (defined below).
Because infinite sequences of binary digits can be identified with real numbers in the unit interval, random binary sequences are often called random real numbers. Additionally, infinite binary sequences correspond to characteristic functions of sets of natural numbers; therefore those sequences might be seen as sets of natural numbers.
The class of all MartinLöf random (binary) sequences is denoted by RAND or MLR.
Contents
History
The first suitable definition of a random sequence was given by Per MartinLöf in 1966. Earlier researchers such as Richard von Mises had attempted to formalize the notion of a test for randomness in order to define a random sequence as one that passed all tests for randomness; however, the precise notion of a randomness test was left vague. MartinLöf's key insight was to use the theory of computation to formally define the notion of a test for randomness. This contrasts with the idea of randomness in probability; in that theory, no particular element of a sample space can be said to be random.
MartinLöf randomness has since been shown to admit many equivalent characterizations — in terms of compression, randomness tests, and gambling — that bear little outward resemblance to the original definition, but each of which satisfy our intuitive notion of properties that random sequences ought to have: random sequences should be incompressible, they should pass statistical tests for randomness, and it should be difficult to make money betting on them. The existence of these multiple definitions of MartinLöf randomness, and the stability of these definitions under different models of computation, give evidence that MartinLöf randomness is a fundamental property of mathematics and not an accident of MartinLöf's particular model. The thesis that the definition of MartinLöf randomness "correctly" captures the intuitive notion of randomness has been called the MartinLöf–Chaitin Thesis; it is somewhat similar to the Church–Turing thesis.^{[1]}
Three equivalent definitions
MartinLöf's original definition of a random sequence was in terms of constructive null covers; he defined a sequence to be random if it is not contained in any such cover. Leonid Levin and ClausPeter Schnorr proved a characterization in terms of Kolmogorov complexity: a sequence is random if there is a uniform bound on the compressibility of its initial segments. Schnorr gave a third equivalent definition in terms of martingales (a type of betting strategy). Li and Vitanyi's book An Introduction to Kolmogorov Complexity and Its Applications is an excellent introduction to these ideas.
 Kolmogorov complexity (Schnorr 1973, Levin 1973): Kolmogorov complexity can be thought of as a lower bound on the algorithmic compressibility of a finite sequence (of characters or binary digits). It assigns to each such sequence w a natural number K(w) that, intuitively, measures the minimum length of a computer program (written in some fixed programming language) that takes no input and will output w when run. Given a natural number c and a sequence w, we say that w is cincompressible if .
 An infinite sequence S is MartinLöf random if and only if there is a constant c such that all of S's finite prefixes are cincompressible.
 Constructive null covers (MartinLöf 1966): This is MartinLöf's original definition. For a finite binary string w we let C_{w} denote the cylinder generated by w. This is the set of all infinite sequences beginning with w, which is a basic open set in Cantor space. The product measure μ(C_{w}) of the cylinder generated by w is defined to be 2^{w}. Every open subset of Cantor space is the union of a countable sequence of disjoint basic open sets, and the measure of an open set is the sum of the measures of any such sequence. An effective open set is an open set that is the union of the sequence of basic open sets determined by a recursively enumerable sequence of binary strings. A constructive null cover or effective measure 0 set is a recursively enumerable sequence U_{i} of effective open sets such that and for each natural number i. Every effective null cover determines a G_{δ} set of measure 0, namely the intersection of the sets U_{i}.
 A sequence is defined to be MartinLöf random if it is not contained in any G_{δ} set determined by a constructive null cover.
 Constructive martingales (Schnorr 1971): A martingale is a function such that, for all finite strings w, , where is the concatenation of the strings a and b. This is called the "fairness condition"; a martingale is viewed as a betting strategy, and the above condition requires that the better plays against fair odds. A martingale d is said to succeed on a sequence S if where is the first n bits of S. A martingale d is constructive (also known as weakly computable, lower semicomputable, subcomputable) if there exists a computable function such that, for all finite binary strings w
 for all positive integers t,
 A sequence is MartinLöf random if and only if no constructive martingale succeeds on it.
 (Note that the definition of martingale used here differs slightly from the one used in probability theory.^{[2]} That definition of martingale has a similar fairness condition, which also states that the expected value after some observation is the same as the value before the observation, given the prior history of observations. The difference is that in probability theory, the prior history of observations just refers to the capital history, whereas here the history refers to the exact sequence of 0s and 1s in the string.)
Interpretations of the definitions
The Kolmogorov complexity characterization conveys the intuition that a random sequence is incompressible: no prefix can be produced by a program much shorter than the prefix.
The null cover characterization conveys the intuition that a random real number should not have any property that is “uncommon”. Each measure 0 set can be thought of as an uncommon property. It is not possible for a sequence to lie in no measure 0 sets, because each onepoint set has measure 0. MartinLöf's idea was to limit the definition to measure 0 sets that are effectively describable; the definition of an effective null cover determines a countable collection of effectively describable measure 0 sets and defines a sequence to be random if it does not lie in any of these particular measure 0 sets. Since the union of a countable collection of measure 0 sets has measure 0, this definition immediately leads to the theorem that there is a measure 1 set of random sequences. Note that if we identify the Cantor space of binary sequences with the interval [0,1] of real numbers, the measure on Cantor space agrees with Lebesgue measure.
The martingale characterization conveys the intuition that no effective procedure should be able to make money betting against a random sequence. A martingale d is a betting strategy. d reads a finite string w and bets money on the next bit. It bets some fraction of its money that the next bit will be 0, and then remainder of its money that the next bit will be 1. d doubles the money it placed on the bit that actually occurred, and it loses the rest. d(w) is the amount of money it has after seeing the string w. Since the bet placed after seeing the string w can be calculated from the values d(w), d(w0), and d(w1), calculating the amount of money it has is equivalent to calculating the bet. The martingale characterization says that no betting strategy implementable by any computer (even in the weak sense of constructive strategies, which are not necessarily computable) can make money betting on a random sequence.
Properties and examples of MartinLöf random sequences
 Chaitin's halting probability Ω is an example of a random sequence.
 RAND^{c} (the complement of RAND) is a measure 0 subset of the set of all infinite sequences. This is implied by the fact that each constructive null cover covers a measure 0 set, there are only countably many constructive null covers, and a countable union of measure 0 sets has measure 0. This implies that RAND is a measure 1 subset of the set of all infinite sequences.
 Every random sequence is normal.
 There is a constructive null cover of RAND^{c}. This means that all effective tests for randomness (that is, constructive null covers) are, in a sense, subsumed by this universal test for randomness, since any sequence that passes this single test for randomness will pass all tests for randomness. (MartinLöf 1966)
 There is a universal constructive martingale d. This martingale is universal in the sense that, given any constructive martingale d, if d succeeds on a sequence, then d succeeds on that sequence as well. Thus, d succeeds on every sequence in RAND^{c} (but, since d is constructive, it succeeds on no sequence in RAND). (Schnorr 1971)
 The class RAND is a subset of Cantor space, where refers to the second level of the arithmetical hierarchy. This is because a sequence S is in RAND if and only if there is some open set in the universal effective null cover that does not contain S; this property can be seen to be definable by a formula.
 There is a random sequence which is , that is, computable relative to an oracle for the Halting problem. (Schnorr 1971) Chaitin's Ω is an example of such a sequence.
 No random sequence is decidable, computably enumerable, or cocomputablyenumerable. Since these correspond to the , , and levels of the arithmetical hierarchy, this means that is the lowest level in the arithmetical hierarchy where random sequences can be found.
 Every sequence is Turing reducible to some random sequence. (Kučera 1985/1989, Gács 1986). Thus there are random sequences of arbitrarily high Turing degree.
Relative randomness
As each of the equivalent definitions of a MartinLöf random sequence is based on what is computable by some Turing machine, one can naturally ask what is computable by a Turing oracle machine. For a fixed oracle A, a sequence B which is not only random but in fact satisfies the equivalent definitions for computability relative to A (e.g., no martingale which is constructive relative to the oracle A succeeds on B) is said to be random relative to A. Two sequences, while themselves random, may contain very similar information, and therefore neither will be random relative to the other. Any time there is a Turing reduction from one sequence to another, the second sequence cannot be random relative to the first, just as computable sequences are themselves nonrandom; in particular, this means that Chaitin's Ω is not random relative to the halting problem.
An important result relating to relative randomness is van Lambalgen's theorem, which states that if C is the sequence composed from A and B by interleaving the first bit of A, the first bit of B, the second bit of A, the second bit of B, and so on, then C is algorithmically random if and only if A is algorithmically random, and B is algorithmically random relative to A. A closely related consequence is that if A and B are both random themselves, then A is random relative to B if and only if B is random relative to A.
Stronger than MartinLöf randomness
Relative randomness gives us the first notion which is stronger than MartinLöf randomness, which is randomness relative to some fixed oracle A. For any oracle, this is at least as strong, and for most oracles, it is strictly stronger, since there will be MartinLöf random sequences which are not random relative to the oracle A. Important oracles often considered are the halting problem, , and the nth jump oracle, , as these oracles are able to answer specific questions which naturally arise. A sequence which is random relative to the oracle is called nrandom; a sequence is 1random, therefore, if and only if it is MartinLöf random. A sequence which is nrandom for every n is called arithmetically random. The nrandom sequences sometimes arise when considering more complicated properties. For example, there are only countably many sets, so one might think that these should be nonrandom. However, the halting probability Ω is and 1random; it is only after 2randomness is reached that it is impossible for a random set to be .
Weaker than MartinLöf randomness
Additionally, there are several notions of randomness which are weaker than MartinLöf randomness. Some of these are weak 1randomness, Schnorr randomness, computable randomness, partial computable randomness. Additionally, KolmogorovLoveland randomness is known to be no stronger than MartinLöf randomness, but it is not known whether it is actually weaker.
See also
 Random sequence
 Gregory Chaitin
 stochastics
 Monte Carlo method
References
 ^ JeanPaul Delahaye, Randomness, Unpredictability and Absence of Order, in Philosophy of Probability, p. 145167, Springer 1993.
 ^ John M. Hitchcock and Jack H. Lutz (2006). "Why computational complexity requires stricter martingales". Theory of Computing Systems.
 Rod Downey, Denis R. Hirschfeldt, Andre Nies, Sebastiaan A. Terwijn (2006). "Calibrating Randomness". The Bulletin of Symbolic Logic 12 (3/4): 411–491. doi:10.2178/bsl/1154698741.
 Gács, P. (1986). "Every sequence is reducible to a random one". Information and Control 70 (2/3): 186–192. doi:10.1016/S00199958(86)800043.
 Kučera, A. (1985). "Measure, Π_{1}^{0}classes and complete extensions of PA". Recursion Theory Week. Lecture Notes in Mathematics 1141, SpringerVerlag. pp. 245–259.
 Kučera, A. (1989). "On the use of diagonally nonrecursive functions". Studies in Logic and the Foundations of Mathematics. 129. NorthHolland. pp. 219–239.
 Levin, L. (1973). "On the notion of a random sequence". Soviet Mathematics Doklady 14: 1413–1416.
 Li, M.; Vitanyi, P. M. B. (1997). An Introduction to Kolmogorov Complexity and its Applications (Second ed.). Berlin: SpringerVerlag.
 MartinLöf, P. (1966). "The definition of random sequences". Information and Control 9 (6): 602–619. doi:10.1016/S00199958(66)800189.
 Schnorr, C. P. (1971). "A unified approach to the definition of a random sequence". Mathematical Systems Theory 5 (3): 246–258. doi:10.1007/BF01694181.
 Schnorr, C. P. (1973). "Process complexity and effective random tests". Journal of Computer and System Sciences 7 (4): 376–388. doi:10.1016/S00220000(73)800303.
 Ville, J. (1939). Etude critique de la notion de collectif. Paris: GauthierVillars.
Categories:
Wikimedia Foundation. 2010.
Look at other dictionaries:
Random sequence — A random sequence is a kind of stochastic process. In short, a random sequence is a sequence of random variables. Random sequences are essential in statistics. The statistical analysis of any experiment usually begins with the words let X 1,...,… … Wikipedia
Random number — may refer to:* A number generated for or part of a set exhibiting statistical randomness. * A random sequence obtained from a stochastic process. * An algorithmically random sequence in algorithmic information theory. * The output of a random… … Wikipedia
Per MartinLöf — in 2004 Born May 8, 1942 (194 … Wikipedia
Algorithmic information theory — is a subfield of information theory and computer science that concerns itself with the relationship between computation and information. According to Gregory Chaitin, it is the result of putting Shannon s information theory and Turing s… … Wikipedia
Reverse mathematics — is a program in mathematical logic that seeks to determine which axioms are required to prove theorems of mathematics. The method can briefly be described as going backwards from the theorems to the axioms. This contrasts with the ordinary… … Wikipedia
Kolmogorov complexity — In algorithmic information theory (a subfield of computer science), the Kolmogorov complexity of an object, such as a piece of text, is a measure of the computational resources needed to specify the object. It is named after Soviet Russian… … Wikipedia
Wikipedia:Artículos solicitados — Atajos WP:AS WP:SOL Artículos solicitados En esta página pue … Wikipedia Español
Wikiproyecto:Matemáticas — … Wikipedia Español
Claus P. Schnorr — Claus Peter Schnorr C. P. Schorr at Oberwolfach, 1986. Born 4 August 1943 … Wikipedia
Normal number — For the floating point meaning in computing, see normal number (computing). In mathematics, a normal number is a real number whose infinite sequence of digits in every base b[1] is distributed uniformly in the sense that each of the b digit… … Wikipedia