Content deleted Content added
m Guanaco - robot: changing unnecessary HTML entities to ISO-8859-1 characters - see Wikipedia:Special characters |
fixed invariance thm, added technicality and changed notation in line with the most common standard. |
||
Line 1:
[[zh:算法信息论]]
'''Algorithmic information theory''' is a field of study which attempts to capture the concept of complexity by using tools from theoretical computer science. The chief idea is to define the complexity (or '''Kolmogorov complexity''') of a [[string]] as the length of the shortest program which
The field was developed by [[Andrey Kolmogorov]], [[Ray Solomonoff]] and [[Gregory Chaitin]] starting in the late [[1960s]]. There are several
Line 7:
To formalize the above definition of complexity, one has to specify exactly what types of programs are allowed. Fortunately, it doesn't really matter: one could take a particular notation for [[Turing machine|Turing machines]], or [[Lisp programming language|LISP]] programs, or [[Pascal programming language|Pascal]] programs, or [[Java virtual machine]] bytecode.
If we agree to measure the lengths of all objects consistently in [[bit|bits]], then the resulting notions of complexity will only differ by a constant :<math>
Here, ''
(One technical requirement is that it must be possible to embed arbitrary binary data into programs without
delimeters, e.g. by providing such data on "standard input" and considering all bits read
from this stream as part of the program.)
In the following, we will fix one definition and simply write ''
The first surprising result is that ''
It is however straightforward to compute upper bounds for ''
The next important result is about the [[randomness]] of strings. Most strings are complex in the sense that they cannot be significantly compressed: ''
the [[probability]] that a random string of length '' :''I will attach a prize of $5,000 to anyone who successfully meets this challenge. First, the contestant will tell me HOW LONG of a data file to generate. Second, I will generate the data file, and send it to the contestant. Last, the contestant will send me a decompressor and a compressed file, which will together total in size less than the original data file, and which will be able to restore the compressed file to the original state.''
Line 23 ⟶ 28:
Now for Chaitin's incompleteness result: though we know that most strings are complex in the above sense, the fact that a specific string is complex can never be proven (if the string's length is above a certain threshold). The precise formalization is as follows. Suppose we fix a particular consistent [[axiomatic system]] for the [[natural number|natural numbers]], say [[Peanos axioms|Peano's axioms]]. Then there exists a constant ''L'' (which only depends on the particular axiomatic system and the choice of definition of complexity) such that there is no string ''s'' for which the statement
:<math>
can be proven within the axiomatic system (even though, as we know, the vast majority of those statements must be true). Again, the proof of this result is a formalization of Berry's paradox.
|