Jump to content

Talk:Kolmogorov complexity

Page contents not supported in other languages.
From Wikipedia, the free encyclopedia


Must description languages be Turing complete?

[edit]

This article does not say whether the description language is required to be Turing complete. But it links to minimum message length, which makes the claim that Kolmogorov complexity allows only Turing complete description languages. I don't know whether that claim is true or not. If that claim is true, then this requirement should be described in the first paragraph that mentions description languages (the existing link to description language is useless, and links to a page listing languages like XML which are not Turing complete). If that claim is not true, then the first theorem here should be fixed, by adding a phrase requiring that L1 and L2 be Turing complete. Clearly, there are cases where the theorem is false, if L1 and L2 don't have to be Turing complete.

It is understood that "the" Kolmogorov complexity is with respect to a universal description system -- i.e. the description language is Turing complete. In the literature there are also definitions of Kolmogorov complexity relative to a particlar (not necessarily universal) description system, but that is not what is generally intended by "the" Kolmogorov complexity. The latter expression is still a slight abuse of terminology, because the KC will still depend on which universal system is used.
--r.e.s. (talk) 19:19, 28 July 2008 (UTC) I've revised the introduction of the article to reflect this. --r.e.s. (talk) 19:28, 28 July 2008 (UTC)[reply]

Spelling

[edit]

My bot semiautomatically fixed a mistake in this page, see [2] doing the replacement

parameterisation ->
parametrisation

I got a an anonimous message today that the former is also valid spelling. Is that true? I myself would think that it is not in this context. Oleg Alexandrov 6 July 2005 15:02 (UTC)

I think your bot spells correctly, although I am an American, and would use a 'z' instead of an 's'. The former spelling looks incorrect to me in any case. -- 130.94.162.61 23:59, 21 February 2006 (UTC)[reply]
See these pages [3] [4] in Merriam-Webster Online Dictionary. :-) CiaPan 01:17, 22 February 2006 (UTC)[reply]

There's no such word as 'parametrisation' the correct spelling is 'parameterization' - source the OED (Oxford English Dictionary) Fustbariclation (talk) 04:02, 17 August 2022 (UTC)[reply]

Opening definition

[edit]

The first sentence reads as follows:

In computer science, algorithmic information theory is a field of study which attempts to capture the concept of complexity by using tools from theoretical computer science.

Typically the first sentence of an article is some definition or a thumbnail sketch of one. The above statement is not. Additionally, it could have equally applied to the better known field of Computational complexity theory.

I plan to fix this as time permits. Vonkje 22:42, 23 July 2005 (UTC)[reply]


Suggestion to move

[edit]

I propose that this page be moved to Kolmogorov complexity. This is the main subject of the article and would simplify the article's introduction.--CSTAR 01:56, 3 October 2005 (UTC)[reply]


Compression

[edit]

"incompressible strings must exist, since there are bit strings of length n but only shorter strings, that is strings of length n − 1."

I'm no expert, but I don't think this makes sense... there must be something like shorter strings (more than for n > 3), and there is no reason to expect a string to be compressed in exactly one less symbol. Can anyone restate this proof more clearly?

Take a set ("an alphabet") of A symbols. There is different N-letter strings (sequences of length N with values from that set). For binary strings (two-symbol set, A=2) there is strings.
There is, of course strings of length N-1, strings of length N-2, strings of length N-3, and so on, down to strings of length 1. These terms add up to the sum known as a finite geometric series:
which is less than the number of N-bit strings.
Actually it may be enough to compare just the number of different strings of length N and N-1, because the same problem arises on every N level — if you want to compress N-bit strings into strings shorter than N-1, then you use some subset of compression function's codomain, that might be occupied by compressed N-1-bit strings. But whichever way you calculate, you will get the same result: for every N>0 there is more strings of length N (or shorter) to be compressed than strings of length N-1 (or shorter), that they might be compressed into.
CiaPan 07:55, 8 February 2006 (UTC) (Hope my explanation is clear enough - I'm not as fluent in English as in maths.) [reply]


What about zero length strings? therefore sum from to



The fact that there are more strings with exactly length N than all of the strings with length < N is not immediately obvious. Although the quoted sentence is correct ie there are only strings of length N-1, the use of the word shorter makes this confusing.
I suggest
"incompressible strings must exist, since there are bit strings of length n but only shorter strings, that is strings of length n − 1 or less."
Yes No?
Spacepickle 16:45, 31 May 2006 (UTC)[reply]


Short answer: If you feel your version is more clear, insert it. Be bold.
Explanation: In the previous notes, when I used '(or shorter)' I meant 'whether you count N-bit and (N-1)-bit strings or all strings with length up to N and up to (N-1) bits, you get the same result'. In other words I tried to merge two sentences in one: 'there is more binary strings with length N than strings with length (N-1), and by recursion, there is more strings with length N or less than strings with length (N-1) or less'.
So it does not matter, whether you consider only strings with given length, or include also all shorter string—there is always more strings to be compressed, than possible results of compression.
More comments: There are three kinds of mathematicians: those who can count, and those who don't...
And among those who can, there are 10 kinds: those who understand binary, and those who don't.
For those who can count and understand binary, the proposition you expressed above—that there is more strings of length N than all strings shorter than N— is immediately obvious.
For length N There is # strings of length
exactly N any, less than N
binary decimal binary decimal
0 1 1 0 0
1 10 2 1 1
2 100 4 11 3
3 1000 8 111 7
4 10000 16 1111 15
... ... ... ... ...
N 1 and N zeros 2N N ones 2N-1
because there is exactly one (1) empty string and no (0) strings shorter than that:
NumEqual( 0 ) = 1
NumShorter( 0 ) = 0
which is displayed by the first row in the table, and for any natural N:
NumEqual( N+1 ) = 2 × NumEqual( N )
NumShorter( N+1 ) = NumShorter( N ) + NumEqual( N ).
And these recurssions resolve to what the last row of table shows. --CiaPan 17:00, 1 June 2006 (UTC)[reply]

--

The derivation of 2n -1 needs a source, or needs to be corrected. The issue is the 0-length string. One author -- Martin Davis in his "What is a computation?" (pages 241-267) in Lynn Arthur Steen 1980 Mathematics Today: Twelve Informal Essays, Vintage Books, NY, ISBN 0-394-74503-5, excludes the 0-length string and derives from the geometric series that there are strings of length less than or equal to n-10:

2 + 4 + ... + 2n-10 = 2n-10 - 2 (cf page 265)

So there are two things wrong in the article: First, because the little factoid is not readily apparent, the geometric series should be presented before the proof; Second, either the fact that the 0-length string is included or (since at least one author excludes it) the formula should be corrected. I vote for the correction of the formula (as noted in the beginning of this section, the matter has come up before). I'll leave this here for a while and then, because it's bugging me, I'll fix it if no one objects. BillWvbailey (talk) 14:32, 2 June 2010 (UTC)[reply]

The 0-length string is an element of and should be included. The standard reference text on Kolmogorov complexity by Li and Vitanyi, and the new book by Downey and Hirschfeldt that will become the new standard reference, both present the same proof of the existence of random strings as here, and both these texts count strings of length less than $n$. It is not surprising that a popularization might exclude the empty string for pedagogical purposes, but this is not done in the actual literature on Kolmogorov complexity. — Carl (CBM · talk) 17:26, 10 June 2010 (UTC)[reply]
Every real (practical, existent, whatever you want to call it) system of information is able to represent empty string – for example "" in C and '' in Pascal programming languages source (a zero-byte alone terminating the empty string and a zero-length field alone in popular implementations, respectively) or a pair of signals StartOfTransmission – EndOfTransmission in any serial communication system. So there is no reason (except possibly someone's phobia) to avoid considering empty string. No such issue actualy exist. Article fixed. [5] --CiaPan (talk) 10:21, 5 January 2011 (UTC)[reply]

Chiapan's explanation is actually rather important. Chaitin discusses this in his [I know, I know ... ] popularization, MetaMath!, and as I recall so did Martin Davis (with an actual example). Should we add something about (the notion of [the problem of]) self-delimited, variable-length information strings (aka "messages")? Perhaps an example? It would help us show-me types, I think. In other words, the string, if it exists, exists between two end-markers "BEGIN" and "END", or is otherwise "delimited". But the difficulty arises because of the strings' (unknown) variable lengths, which could be 0. The trivial (upper-value) solution is similar to a biphase modulation serial communication link: e.g. the bit string 11 indicates the BEGIN and END markers, 10 = "value = 0", 01 = "value = 1", 00 = blank tape, empty channel, no transmission. So the smallest string 1111 indicates the "empty string" on an otherwise-blank tape. This is an "expensive" solution (it requires 2x bandwidth i.e. 2*N+4 tape-squares). But it does represent the "upper value" number of squares (bits) required to encode any message of length N. Another approach is to start with a "preamble" that contains, e.g. in biphase-encoded binary, the number N of binary string-bits to follow. So no end-marker is required, technically -- the Turing machine's program has to count the number of bits in the string. But always there will be a preamble C, in this case C=4+2*log2(N), with a minimum length e.g. 4 [ 1111 ] (so the total message-length would be N+C, where C = 4+2*log2(N)). [O.R. alert: I'd be willing to wager that there exists (and this is it) a theoretically-shortest preamble length, ergo a shortest uncompressed message length.] BillWvbailey (talk) 16:12, 5 January 2011 (UTC)[reply]

The point about there necessarily being one uncompressed string to which a compressed string must map seems to be dependent on the assumption that we are working with a fixed universal Turning machine. If we expand our view to allow for multiple universal Turing machines, then it seems that a compressed string could map to more than one uncompressed string and furthermore allow for the possibility that a string that is incompressible giving one UTM is compressible given another. This situation appears to have a very interesting connection to the "No Free Lunch Theorem" in machine learning, with the UTM representing assumptions or biases required to make induction possible. Someone who has a better grasp of the math should confirm that this connection exists and add it to the main page, or at least point out that the application of the pigeonhole principle in this section assumes a fixed UTM. TLP (talk) 17:00, 27 August 2013 (UTC)[reply]

We must be operating with a fixed TM (not necessarily a UTM). To bring this to absurdity, we could chose a different TM for each target, with all mapping from the 0-length string. — Arthur Rubin (talk) 02:42, 7 December 2013 (UTC)[reply]
If you could losslessly compress any string of numbers into something shorter, then you could compress the compressed result and then compress that. Ultimately, you'd be able to compress every possible number down to a single bit...clearly since a single bit can only have two states, it can only code for two different numbers - so this must be impossible. SteveBaker (talk) 03:24, 28 November 2015 (UTC)[reply]

Some issues

[edit]

Why does Talk:Algorithmic information theory forward to Talk:Kolmogorov complexity?

Fixed. The following comment copied over to appropriate talk page.--CSTAR 17:31, 17 June 2006 (UTC)[reply]

The following remarks pertain to the article Algorithmic Information Theory. The article as it stands has some issues that I would like to bring to discussion. I hope that some consensus can be reached about how to deal with them.

  • There are no references.
  • Komogorov complexity is an active research area in mathematics as well as computer science.
  • The statement
Unlike classical information theory, algorithmic information theory gives formal,
rigorous definitions of a random string.....

is a point of view that is not universally shared, although it has been championed by Chaitin in popularizations of the area. It is well known that the AIT definition of random string is actually only asymptotically well defined, so the viewpoint that AIT does not give a formal definition of a random string is equally valid (and, among mathematicians I have spoken with, a common viewpoint). The issue is the lack of a canonical choice of a universal machine.

  • The definition of random infinite sequence has the same extent as the one given by Martin Lof, and this should be mentioned.

CMummert 04:37, 16 June 2006 (UTC)[reply]

Example strings

[edit]

"The first string admits a short English language description namely "50 repetitions of '01'", which consists of 20 characters. The second one has no obvious simple description other than writing down the string itself, which has 100 characters." <----- This is not true. One could reference this string as eg. "the uncompressable 100 character string in http://wiki.riteme.site/wiki/Kolmogorov_complexity" which would have 95 characters.

What do you wiki-authors do with this now? I didn't dare to put it directly in the main page and wanted to spur some discussion. As far as I remember Martin Garder is talking about this paradox, but I cannot find the source right now. (Markus Gaelli)

This is just a red herring. The paradox is just because Kolmogorov complexity is only defined relative to a particular prefix free universal machine. There are plenty of machines that make the repeated 01 string have higher Kolmogorov complexity than the other string, by carefully choosing indices. So the correct reading of the intro paragraph is to view it as an informal motivating example, which is fine for an intro paragraph. Once the formal definition is given, it becomes clear that the intro paragraph is speaking of some intuitive idea of description which is made formal via universal machines. CMummert 23:57, 13 August 2006 (UTC)[reply]

Mandelbrot set

[edit]

The picture is very nice, but I don't see what this has to do with Kolmogorov complexity, since the object in question is not encoded by a finite sequence in any obvious way. Could you provide a reference?--CSTAR 02:02, 24 October 2006 (UTC)[reply]

If you look at the article for Mandelbrot set, it should be clear that this is encodable by a short program. The information theory textbook:
  • Thomas M. Cover, Joy A. Thomas. Elements of information theory, 1st Edition. New York: Wiley-Interscience, 1991. ISBN 0-471-06259-6.
2nd Edition. New York: Wiley-Interscience, 2006. ISBN 0-471-24195-4.
uses such a picture for its front cover and makes a similar point. Calbaer 18:43, 24 October 2006 (UTC)[reply]
I think CSTAR's point is that Kolmogorov complexity is defined for finite strings, not for subsets of the plane. Talking about the Kolmogorov complexity of the Mandelbrot set is, on its face, a category error. The Kolmogorov complexity of a program that generates the Mandelbrot set is another matter. Though even there, the claim that the complexity is "close to zero" is confusing. Is a thousand close to zero? I doubt its complexity is less than that, using a standard definition of universal Turing machine.
Since the Mandelbrot set is closed, you can code it naturally by a single real. It's probably true that the initial segments of that real have bounded Kolmogorov complexity. Is that what you meant? --Trovatore 20:21, 4 November 2006 (UTC)[reply]
The question of what it means for a real to code the Mandelbrot set doesn't have a unique answer. The Mandelbrot set can be coded via a computable enumeration of all the basic open balls in its complement (the fact that there is a computable sequence to do this is not obvious but has been proved). The Kolmogorov complexity of initial segments of a real can never be bounded, but will be bounded by when x is a computable real. CMummert 19:40, 8 November 2006 (UTC)[reply]
Well, if you think it's confusing, it need not be there. However, I adjusted the caption to hopefully make it more understandable and accurate. By the way, back in the 80s, I made some programs to create 640x480 bitmaps of fractals; on a PC, they compiled to approximately 2500 bits. Compared to most image-creating programs, that's "almost zero," but if there's a better way to phrase the caption, I've no objection to its being rephrased. Calbaer 03:59, 5 November 2006 (UTC)[reply]
OK this assertion is quite different; you're talking about bitmaps of fixed size n of which there are 2n. Note also that in your caption, f really isn't a function on Cn, but a function on some floating point type, elements of which are also representable as finite bitstrings. To say "small" you would have to make an asymptotic assertion as n -> ∞ I suppose this is true, although I dont havce a reference or a proof.--CSTAR 17:20, 8 November 2006 (UTC)[reply]
The following question is open: whether for each n you can (in a uniform way) compute a bitmap with resolution of black, white, and undefined pixels such that any pixel entirely contained in the Mandelbrot set is black and any pixel entirely outside is white. This is hinted at in Mandelbrot set#Further results. There is a paper here [6] CMummert 19:40, 8 November 2006 (UTC)[reply]
The current caption is nice, but is inaccurate, since all versions of the picture are JPEGs. Where's this 17 KB PNG? Unless we replace the JPEGs with PNGs, the caption should be reverted. Calbaer 20:13, 6 December 2006 (UTC)[reply]
You're right - it's a JPEG. I fixed the caption. CMummert 20:24, 6 December 2006 (UTC)[reply]
I still find it problematic. The underlying picture itself is 189 KB, not 17 KB. Also, to create it would require a JPEG encoder in addition to the fractal generator. While I'm sure it could be generated in under 17 KB, this isn't obvious, and, moreover, gzip can reduce its size (by about 1.8%) so saying the complexity is merely "much less" is saying very little. (And, yes, I realize my initial caption had the same problem.) What would be great is if someone could contrast a significant difference between size of code with size of (losslessly compressed) file. (If someone has a DOS emulator handy, I can sent them my 80s-era Mandelbrot code to create a 640x480 image on screen.) But, as it is, things might be a bit confusing. Calbaer 21:12, 6 December 2006 (UTC)[reply]
I wouldn't mind it if the image were removed, and User:CSTAR has made similar comments. The new caption is an attempt to find an informally correct way to keep the image while relating it to the article material.
The fundamental question is whether it is possible to make a caption that is technically correct and relevant to the article. There are more serious technical issues with the current one than you mentioned, namely that Kolmogorov complexity is only defined up to a constant and so, relative to certain universal machines, the bitmap is a random string. The caption I wrote today was only intended to make the image fit into the context of the article - the previous caption gave little indication of why the image was there. Feel free to edit it. CMummert 21:21, 6 December 2006 (UTC)[reply]
This exemple does not describe kolmogorov complexity. The image has a finite size in bits. A fractal contain infinite information in the same domain. There is a loss of information, whe cannot reconstruct the fractal from the image, so the objects represented are not the same. —Preceding unsigned comment added by 81.67.80.174 (talk) 09:22, 8 June 2008 (UTC)[reply]
JPEG seems inappropriate here for another reason: it's lossy. You could reduce the size of the image down to less than 1KB by reducing the quality of the image, but that doesn't really say anything about the complexity of the Mandelbrot set. -LesPaul75 (talk) 17:11, 9 June 2008 (UTC)[reply]
I think the important thing is that kolmogorov complexity is defined in for a given "descriptive language", and that language can certainly have a save_to_file_as_jpeg() function. The original mandelbrot set is defined with a very simple equation/algorithm, and all the stuff to make it pretty and compact the picture doesn't add very much to the small kolmogorov complexity. The picture *looks* complex, but it isn't. The bit string example also used at the beginning of the article, on the other hand, can't be significantly compressed short of the descriptive language having a function that returns that exact string. However, then there will always be other random bit string that can't be so shortened, since the descriptive language has to be finite (by definition) and the language would fill up with longer and longer function names. Wrs1864 (talk) 20:16, 9 June 2008 (UTC)[reply]
To avoid the technical issue of specifying a description language that can deal with JPEG decoding (most concrete description languages in theory are based on Turing machines) I've rewritten the caption to talk about the size of the raw image file. We are not presenting this file for download, but it obviously exists, and it obviously can be generated by a much smaller program in a typical general-purpose description language. I'm still eliding the technical detail that the specific image we offer contains JPEG artifacts that cannot necessarily be duplicated by a program using only the definition of the Mandelbrot set; but I get around this by claiming that this image is "just a picture of" the bitmap in question, rather than literally the bitmap in question. Dcoetzee 20:06, 21 June 2008 (UTC)[reply]
Further study of the Mandelbrot set shows that the set may not be computable; there is a program which produces something close to the Mandelbrot set, with the "distance" approaching 0 as the number of iterations approaches infinity (assuming the interior of the boundary is empty, which is not true of all fractals), but without a good estimate of the "distance" at any time. Perhaps the Koch snowflake, or some other simple fractal, might be better. The Sierpinski carpet is easier to draw, but it doesn't look complicated. — Arthur Rubin (talk) 07:01, 11 August 2014 (UTC)[reply]


Sorry to re-open this discussion again, but the image and its caption aren't yet convincing:

  • The size of the commons original png image is 1'691'464 Bytes = 13'531'712 Bits, not "1.62 million bits" as the caption claims (this is easy to fix).
  • According to the commons description, it was "created with Fractal Forge 2.8.2"; downloading from https://fractal-forge.en.softonic.com and unpacking shows that "FractalForge.exe" has a size of 936'448 Bytes - not impressively smaller than the png image.
  • Assuming it to be written in C, I guess its source code has a fairly small size (when the typical gui fuss is ignored); however, the size of the C compiler (953'040 Bytes for Linux gcc-6, not counting dynamic libraries etc.) should be added to that.
  • The previous considerations presuppose an x86-64 machine instead of a Turing machine. While I think this can be justified, presupposing a C interpreter instead (in order to save the 1 MBytes gcc code) would be somewhat unfair.
  • Possibly, we could just find a much larger image file; the largest I found below commons:Category:Mandelbrot sets is File:Colourful Mandelbrot.jpg with a size of 100'514'793 Bytes; when converted to an uncompressed tif, it gets 240'025'632 Bytes large (= width 8796 pixels * height 9096 pixels * 3 Bytes rgb values per pixel + 384 Bytes header). Assuming this tif image could have been generated by FractalForge.exe gives a compression rate of 240'025'632 / 936'448 = 256.3.
  • To get a really impressing rate, we could use a program in a very simple language, i.e. with a very small compiler (or interpreter); like Basic (programming language). This could also illustrate that computation time is not an issue in Kolmogorov complexity theory. Probably, the language needs to support floating-point arithmetic (not sure that Basic does). Maybe an ancient C compiler would be sufficiently small, too, for our purpose?

Sorry to bother with these pedantic thoughts, but I think if we give exact figures in the example, they should be correct and illustrative. And I think a Mandelbrot image is a good example for a large (and seemingly complex) object generatable by a much smaller program, thus illustrating the core concept below Kolmogorov complexity. - Jochen Burghardt (talk) 18:00, 27 August 2017 (UTC)[reply]

I'd agree that the Mandelbrot image is an example of a complex structure arising from the application of conceptually simple rules; but shouldn't it be self-evident how a motivating example in an encyclopaedia article relates to the topic?
As evinced by this discussion, you have to work quite hard to see how what looks like a complicated subset of the complex plane relates to a property of finite strings. (Or for that matter how a relatively simple predicate in a mathematical context translates to the length of a computer program for generating the set of points for which the predicate holds.)
How about a different, less controversial example, perhaps Langton's ant? NeilOnWiki (talk) 15:56, 5 March 2018 (UTC)[reply]

Uncomputability of Kolmogorov complexity

[edit]

It seems to me that the program GenerateParadoxicalString is completely unnecessary in the proof, as the desired contradiction is immediate from GenerateComplexString itself. All that's needed is to note the following contradictory consequences:

(1) GenerateComplexString, with input n, returns a string whose Kolmogorov complexity is at least n; consequently, for all n the combined length of GenerateComplexString and n, is at least n.

(2) The combined length of GenerateComplexString and n, is just a constant plus the length of a numeral for n; consequently, for sufficiently large n this combined length is strictly less than n. --r.e.s. 05:37, 11 February 2007 (UTC)[reply]

The reason for the GenerateParadoxicalString function is to make it clear that there really is a "sufficiently large" value of n that works. Also, you need to be careful in (1) because adding two strings together can give you a string with much lower Kolmogorov complexity than the sum of the complexities of the strings. CMummert · talk 13:25, 11 February 2007 (UTC)[reply]
Can't the point concerning "sufficiently large n" be made just as clearly using GenerateComplexString itself, rather than GenerateParadoxicalString? Also, doesn't the caution about Kolmogorov complexity when "adding" two strings apply with equal force to GenerateParadoxicalString? If so, the suggested change would yield a proof that's not only clearer and more direct, but also shorter by about half. --r.e.s. 14:24, 11 February 2007 (UTC)[reply]
Feel free to rewrite the proof; if there is something wrong with your revised version, either I or someone else will point it out. My comment above points out the two main areas where you need to take caution in your rewrite. If you want to be more cautious, you can put a new version here on the talk page for others to comment on. CMummert · talk 15:17, 11 February 2007 (UTC)[reply]
It doesn't make sense to rewrite it only to have the logic disputed afterwards, which is why I asked the two questions above (they were not merely rhetorical). I would prefer to discuss the basis before building on it. However, another fly in the ointment is the likelihood that someone will soon come along arguing that these "original" proofs don't belong in Wikipedia (and they'd be right, I guess, as I don't see any source cited for the present proof either), and all effort will have been wasted. Maybe the wiser course (for someone energetic) would be to locate an authoritative source and just cite it, paraphrasing the proof only if it's very short. --r.e.s. 16:36, 11 February 2007 (UTC)[reply]
Reply to r.e.s: Your proof sketch seems OK to me, although note that all the proofs in this page are informal, and when written down using a formalization of string descriptions (using Turing machines,) I'm not sure there is going to be any substantive difference in clarity or length. Also, for an expository and informal article, brevity of proofs is not necessarilly always better (although it usually is). --CSTAR 17:21, 11 February 2007 (UTC)[reply]
The proof in this article (which is really the same as what you sugggested above, just phrased differently) is not "original" in the WP sense; a proof of the result in question following the same lines could be found in Li and Vitanyi's book or Calude's book. There is broad consensus in mathematics articles that it is acceptable to include proofs of well-known results when those proofs are useful to explain what is going on. The proof we are discussing here is exactly that sort of proof. So there is no reason to remove the proof entirely.
The two bullets you gave above are not enough on their own to consitute a proof for an intended reader of this article, which is why I can't say what else is "needed" for an alternate proof. I just pointed out what the touchy points will be. CMummert · talk 18:45, 11 February 2007 (UTC)[reply]
I disagree that (1) and (2) are not enough, in that the proposed proof is at least as solid as the one in the article. Concerning your word of caution on the complexity of two "added" strings ... The two strings GenerateComplexString and its input n together constitute the program that's supposed to return a string with the property that all programs returning it have length at least n. (Taking the input as part of the program in this case is consistent with the way it's done more formally with TMs; e.g., as the article puts it, "if M is a TM which on input w outputs string x, then the concatenated string <M> w is a description of x".) Thus the inequality in (1) must hold by the very definition of Kolmogorov complexity. On the other hand, to question bullet (2) is to question an argument that's already used in the article, though applied there to the obfuscating program GenerateParadoxicalString.
However, I'll leave the suggested improvement to others if they wish to pursue it, as I doubt whether I have the energy to dispute the matter much further. --r.e.s. 02:04, 12 February 2007 (UTC)[reply]
The difficulty with the argument you just presented is: what is the value of n that actually leads to a contradiction? In order to turn an input into a program, you have to have a particular value of the input to encode (in binary, for example). The purpose of GenerateParadoxicalString is to obtain a bound that lets you choose a specific value of n. CMummert · talk 02:55, 12 February 2007 (UTC)[reply]
There is no difficulty, since all that's needed is that the inequality in (1) holds for all n, and the inequality in (2) holds for some n. Throughout, it's understood that some encoding has been chosen for all inputs n, and that it's the length of the encoding that's meant by "the length of n". At the same level of detail and/or informality as in the current proof, one can establish (1) and (2), and hence the desired contradiction -- with no need of GenerateParadoxicalString.--r.e.s. 12:29, 12 February 2007 (UTC)[reply]
{{sofixit}}. That's what I said above, as well. CMummert · talk 12:55, 12 February 2007 (UTC)[reply]
Thanks for the advice to feel free about doing a rewrite. At this point my choice is to decline.--r.e.s. 13:59, 12 February 2007 (UTC)[reply]

Two kinds of Kolmogorov complexity?

[edit]

Here it says, if I understand correctly, that there are two different definitions of Kolmogorov complexity (plain and prefix). It would be nice to be able to read about that in the article. --Tgr 12:21, 18 September 2007 (UTC)[reply]

I don't believe, as claimed, that the leading texts distinguish between these two definitions, which I've never seen before reading this blog post. In the classroom, generally one definition is given, so that should suffice here, barring a reliable source to the contrary. I suspect that "prefix complexity" means that the program includes information about how long it is, whereas "plain complexity" would not include this information. This is not generally an issue. I believe "plain complexity" is usual definition. However, although the program description is generally not assumed to include a termination indicator, the program contains information about when to terminate during its runtime. Calbaer 00:34, 19 September 2007 (UTC)[reply]
Yes, there is plain Kolmogorov complexity and prefix-free complexity. The latter is crucial for things like Chaitin's omega. There is some description at Chaitin's constant. This article only covers the regular Kolmogorov complexity. — Carl (CBM · talk) 02:48, 19 September 2007 (UTC)[reply]

Example in the intro

[edit]

The English phrase "The binary representation of 14,439,066,989,718,820,758" has less than 64 characters. Army1987 (talk) 00:19, 21 June 2008 (UTC)[reply]

Fixed. The new string was made of the lowest 7 bits of bytes from /dev/random, discarding those < 32 or = 127. So, it's very unlikely to be described in less than 64 characters. --Army1987 (talk) 11:05, 21 June 2008 (UTC)[reply]
I think it would be preferable to rephrase the introductory example in terms of bits, because "character" is so poorly specified (after all, if we're talking Unicode characters, you could probably encode even your random string in half the space). I'm going to attempt this. Dcoetzee 19:22, 21 June 2008 (UTC)[reply]
In the end I decided binary strings were infeasible because it's really hard to describe anything in 64 bits. So I went with specifying the alphabet more carefully: only lowercase letters, numbers, and spaces are allowed. This ensures that the second example is likely to be an incompressible string, while still providing enough characters for English text. Dcoetzee 19:47, 21 June 2008 (UTC)[reply]

Incomputible?

[edit]

But, what if there was a program that counts in binary and executes the result until it outputs the given string?

int getComplexity(string s){ 
string output;
int number=0;
program prog;

 while(1){
 prog = number.toBinary();
 prog.execute();
 output = prog.output;
 if(output==s){break;}
 number++;
 }

return length(number.toBinary());
}

Since this goes through all possible binary values, if there exists a program which outputs the string, it will find it. It will take an enormous amount of processing power, of course. —Preceding unsigned comment added by 69.208.73.205 (talk) 17:20, 10 July 2008 (UTC)[reply]

Well, let's assume your "program" class is carefully constructed so that this thing doesn't just immediately coredump. Then what will happen is, before long, you'll hit a program that never terminates, and then you'll never get to the next value of "number". Nor can you add a program::checkTerminate() method that will figure this out before you call execute(); this is by the unsolvability of the halting problem. --Trovatore (talk) 17:36, 10 July 2008 (UTC)[reply]
As written, the program won't work for the reasons stated by Trovatore; however, it is possible to simulate all possible programs simultaneously by simulating one step of the first program in lexicographical order, then one step of each of the first 2 programs, then one step of the first 3 programs, and so on. The problem with this is that you wouldn't know when to stop running the simulation; it may so happen that a longer program produces the string and halts before a shorter program has yet produced any output. Dcoetzee 20:52, 28 July 2008 (UTC)[reply]
Yes, nicely put. Of course there are infinitely many ways you might try to fix the approach, none of which will work, but this is a reasonably natural one to try. --Trovatore (talk) 21:48, 28 July 2008 (UTC)[reply]
Hold on. If the longer program produces the string then either A) there are no shorter programs running that haven't halted or produced incorrect results in which case this program is the shortest program for the string and you have computed the Kolmogorov complexity, or B) there are shorter programs still running which could in theory produce the string in which case these programs are the lower bound on the Kolmogorov complexity in the sense that no other program not longer than these can produce the string in fewer steps. In either case you can halt then and there because you've contradicted the claim from the introduction that even the lower bounds are uncomputable. 37.223.147.242 (talk) 03:45, 29 February 2020 (UTC)[reply]
That's right; the introduction sentence is awful nonsense (introduced by me in 2017). Of course, e.g. zero is a lower bound for the Kolmogorov complexity, of any string. In fact, Chaitin's incompleteness theorem states that "K(s) ≥ L" can be proven only for L up to a fixed value that depends only on the employed proof calculus, no matter how long the string s is.
I inserted the word "nontrivial" in the lead sentence now; I hope that is sufficient.
Thanks for noticing, and sorry for the confusion. - Jochen Burghardt (talk) 17:34, 29 February 2020 (UTC)[reply]

Question on Vertical Bars in the Equations

[edit]

In these two equations,

and

what do the vertical bars mean in each? It looked like to me that maybe they meant "length" in the first and "absolute value" in the second, but I wasn't sure.

Thank you,

Cat Dancer WS (talk) 14:09, 3 August 2008 (UTC)[reply]

You are correct, on both counts. Dcoetzee 21:04, 4 August 2008 (UTC)[reply]
In the same proof, can you explain how
follows from
Thank you, 91.140.75.204 (talk) 17:37, 30 July 2009 (UTC)[reply]
It doesn't. It follows from
and
That's what the "by symmetry" means. —Preceding unsigned comment added by 130.195.86.38 (talk) 04:17, 16 October 2009 (UTC)[reply]

Kolmogorov Complexity & Information

[edit]

Dear all,

I’m rather a fan of Kolmogorov complexity, than an expert. I believe however that Kolmogorov was on the right track to mathematically describe information and its impact. Unfortunately, there is so much confusion about what is information, especially from IT perspective. And this confusion is reflected in very messy Information article.

I have outlined how the article should present the concept on the related discussion page, but without much of an echo. If there are mathematicians among you who would like to get involved, they will be welcomed.

Kind regards, Damir Ibrisimovic (talk) 04:15, 18 February 2009 (UTC)[reply]

In pursuit of exactness, mathematicians very often hijack Wikipedia articles away from being understandable or usable by those seeking to understand the topic. Ywaz (talk) 23:23, 19 January 2016 (UTC)[reply]

Complexity Density or Entropy Density Comment (tentative)

[edit]

What happens if which divide the Kolmogorov Complexity by the length of the digit, ie : K(s)/|s|, where |s| is the size of s (note, it may be that an appropriate notion of entropy density would be arrived at by dividing by some function of s, or by another complexity metric aside from the Kolmogorov complexity). Do we find that, in some sense, there is a theoretical limitation to the entropy density which we can obtain? Does this not provide a theoretical limitation upon the degree of complexity which can be attained by both formal systems and also applied systems? Is the actual maximum entropy density attainable related to the Chaitin constant?


ConcernedScientist (talk) 18:19, 28 February 2009 (UTC)[reply]

Binary Lambda Calculus

[edit]

I added a link to my recently created page about Binary Lambda Calculus, which provides a relatively simple and easy to work with concrete definition of both plain and prefix complexity.

--Tromp (talk) 14:50, 13 March 2009 (UTC)[reply]

Problem with main example

[edit]

the article suggests that the 64 characters "4c1j5b2p0cv4w1 8rx2y39umgw5q85s7ur qbjfdppa0q7nieieqe9noc4cvafzf" is a description of itself, which is highly misleading. There absolutely needs to be an indication that the string is to be taken literally.

How else are we going to describe the string "ab 32 times" ?

Here we also get into issues with quoting. How do we describe (in English) the string "\"ab 32 times\"" ? Probably the best thing is to avoid character escapes and use a description like "the 13 character string "ab 32 times""

Thus, I suggest the example in the article should have description: "the 64 character string 4c1j5b2p0cv4w1 8rx2y39umgw5q85s7ur qbjfdppa0q7nieieqe9noc4cvafzf"

--216.223.242.211 (talk) 17:28, 13 March 2009 (UTC)[reply]

I see your point, but I don't think it that much of a problem. While the example can be made more accurate with a more formal "description language" than English, the example works to get the point across to people who don't really understand this subject. Wrs1864 (talk) 10:59, 16 March 2009 (UTC)[reply]

Incomputability of Kolmogorov complexity: termination of GenerateComplexString

[edit]

Hello, the function GenerateComplexString(int n) will only terminate for any n if K has no upper bound. Of course, there is no upper bound for K, because if there was, every string (out of about aleph-0 strings) could be generated by a program with a length less than or equal to that bound (of which only a finite number exist). While the lack of proof for the termination of the function is thus just a technicality, it nevertheless requires the reader to think a bit on his own (or, more likely, ignore it). As I am no expert on the topic and would rather not complicate the proof due to a misunderstanding, I will ignore WP:BB and just wait for an expert to verify or refute my concerns. X127 (talk) 07:09, 16 March 2009 (UTC)[reply]

I think the Pigeonhole principle is pretty obvious, but I guess in the top of the "basic results" section where it discusses the maximum length of K(s), something could be added about this. Wrs1864 (talk) 11:18, 16 March 2009 (UTC)[reply]

HOW Incomputable is Kolmogorov complexity/HOW undecidable is the issue of determining the Kolmogorov complexity?

[edit]

This may sound a silly point. And I may be well off the mark in my understanding of what the 'Incomputability of Kolmogorov complexity' entails. BUT, HOW incomputable is Kolmogorov complexity? According to the article Oracle machine, even undecidable problems can be solved by an oracle based equipped Turing machine. So what is the minimal complexity class of decision problems which an oracle attached to a Turing machine (for the purpose of computing the Kolmogorov complexity of numbers) must be capable of solving?

As a separate point, do there exist Polynomial-time approximation schemes which can determine approximations to the Kolmogorov complexity to within an arbitrary constant value of the actual Kolmogorov complexity (it would seem clear that the answer is no in the sense that having a computed Kolmogorov complexity value within 1% of the actual value would enable the Kolmogorov complexity to be calculated - however, some type of probabilistic polynomial-time randomized approximation scheme might enable Kolmogorov complexity evaluation to some probabilistic bound of the 'answer').

Are there any fundamental problems with any of this?

ConcernedScientist (talk) 18:19, 4 April 2009 (UTC)[reply]

First off, as per WP:NOTFORUM, talk pages are supposed to be for discussing the article itself, not for general discussions on the subject matter. As far as what can be computed/proved with atypical systems, I think they would be best discussed on those article pages. E.g., something on Kolmogorov complexity on an Oracle machine, would be best discussed on the Oracle machine page. (The same goes for all other proofs that may change when you can assume you have an oracle.) As far as approximations, as discussed in the article, all strings have a Kolmogorov complexity that is at most the length of the string plus a (small) constant, so yes, there is an obvious polynomial approximation. Wrs1864 (talk) 19:26, 4 April 2009 (UTC)[reply]

1) Oracle machines by definition can solve the halting problem for Turing machines in one operation. There is an infinite hierarchy of them: see Turing degree. The notion of Kolmogorov complexity can be generalized to machines with oracles; algorithmically random sequence has some info. 2) There is no general way (without an oracle) to approximate the Kolmogorov complexity of a string in finite time (much less in P-time). Example: saying "string S has complexity 1 million bits +/- 5%" asserts that none of the programs shorter than 950,000 bits long generate S. But that is a conjunction of a lot of assertions that are undecidable by Rice's theorem. 69.228.170.24 (talk) 20:46, 31 May 2010 (UTC)[reply]


The original poster's question has a precise answer that no one seems to have given: The degree of unsolvability of the problem of computing (non-relativized) Kolmogorov complexity is precisely 0' (read zero-jump), the Turing degree of the halting problem itself. At least I would be very surprised if that were not true (I haven't thought it through at the level of detail that would give a proof). --Trovatore (talk) 21:19, 31 May 2010 (UTC)[reply]

Yes, it is true. Actually, Ω wtt-computes the halting problem [7]. — Carl (CBM · talk) 04:11, 1 June 2010 (UTC)[reply]

what is a "description language"?

[edit]

Do we need a more formal definition/specification of what qualifies as a description language in this article? Right now, the article is pretty vague, just using a link to the vague description language article. Say we have a description language that consists of only Hollerith constants. We can generate any string, but we also do such things as calculate the exact Kolmogorov complexity for each string (the length string, plus one, plus the log base 10 of the string). I thought about throwing in a "turning complete" requirement, but realized that maybe the proofs don't require that powerful of a language, for example, a Linear bounded automaton might still work. Thoughts? Wrs1864 (talk) 02:38, 5 April 2009 (UTC)[reply]

I think one normally defines Kolmogorov complexity relative to a fixed universal Turing machine U. That is, first choose U, and then for a given string s, KU(s), the complexity of s, is the shortest program for U that generates s. Yes, this should be explained more precisely in the article. 69.228.170.24 (talk) 20:36, 31 May 2010 (UTC)[reply]

Chaitin's incompleteness theorem and "minimum message length" ?

[edit]

I don't see why MML is refered to in this section. The notation of MML there is so vague it shows nothing about the relation with CIT, I almost erased it, but I'd rather put it in another section for now. --210.231.57.36 (talk) 12:11, 19 July 2009 (UTC)[reply]

The Wallace/Dowe paper[8] explains it. 69.228.170.24 (talk) 16:42, 29 May 2010 (UTC)[reply]

The Invariance theorem article is a stub with little promise of expansion. The material could easily be merged here. Also, the term "Invariance theorem" pretty generic and could refer to several theorems in a variety of area. It seem to me that once the merge is complete then the existing Invariance theorem article should be changed to either a redirect to Invariant (mathematics) or a dab page.--RDBury (talk) 22:14, 12 July 2010 (UTC)[reply]

  • Yes, merge. My little mind cannot imagine how that article can be expanded beyond what it already says, which is about the same as what this article says. linas (talk) 16:56, 6 June 2012 (UTC)[reply]

The "Theorem" in the Definition section

[edit]

"Theorem. If K1 and K2 are the complexity functions relative to description languages L1 and L2, then there is a constant c (which depends only on the languages L1 and L2) such that

"

I do not understand how this can be true. The problem is that it is always possible to create a sufficiently stupid, but Turing-complete, language that the minimal program for generating a particular string grows more than linearly with the string length in the worst case. An example of such a stupid language would be one in which all elements in the alphabet are equal, but context-dependent, and the only possibility there is for encoding anything lies in the length of the program (Unary expansion of BrainFuck is an example), such that program size grows exponentially with the length of s. At the same time, a language can always be created which only grows linearly with the length of s. This will violate the above theorem.

I suppose, with added conditions, that a similar real theorem exists, and I would be interested in knowing how it should be.

Another question which discloses a rather obvious limitation of this article, is: "What interprets the interpreter program?" (not to say "What interprets the interpreter of the interpreter?", which looks like a vicious regress.) It seems as though some objective basis is needed at the bottom, or at least a real explanation of why the choice of basis does not matter much. —Preceding unsigned comment added by 129.240.72.41 (talk) 14:07, 6 October 2010 (UTC)[reply]

The "proof" has the key: That there is a program in L1 which interprets L2.
A problem with your "language" is that we may no longer have K(s) ≤ |s| + c, (under Basic results), which is generally considered a requirement for a complexity. — Arthur Rubin (talk) 14:15, 6 October 2010 (UTC)[reply]
Stupid question: Isn't the proof not a proof? It supposes the existence of a program in L1 which interprets L2, but never proves that such a program necessarily exists. 66.102.14.9 (talk) 22:06, 7 July 2011 (UTC)[reply]
Since L1 is supposed to be Turing-Complete, then we can assume such a program exists by Church-Turing Thesis.186.206.10.100 (talk) 16:35, 27 January 2014 (UTC)[reply]

Shannon entropy

[edit]

What is the relation of Kolmogorov complexity to Information theory? It seems very close to Shannon entropy, in that both are maximised by randomness, although one seems to deal more with the message than the context. Cesiumfrog (talk) 00:03, 14 October 2010 (UTC)[reply]

Shannon entropy is a statistical measure applied to data which shows how efficiently the symbols are transferring bits based solely on how many times the symbols occur relative to each other. It's like the dumbest, simplest way of looking for patterns in data, which is ironically why it is useful. Its biggest most basic use is for taking data with a large number of different symbols and stating how many bits are needed to say the same thing without doing any compression on the data.
Kolmogorov complexity is the absolute extreme in the direction Shannon entropy takes step 0. It is the highest possible level of intelligent compression of a data set. It is not computable in most cases, it's just a theoretical idea. But Shannon entropy is always computable and blind. By "compression" I mean an algorithm has been added to the data, and the data reduced to a much smaller set that is the input to the algorithm to generate the original data. Kolmogorov is the combination in bits of the program plus the smaller data set.
I would use Shannon entropy to help determine the "real" length in bits of a program that is proposing to be better than others at get close to the idealized Kolmogorov complexity. Competing programs might be using a different or larger sets of functions, which I would assign a symbol to. Programs that use a larger set of functions would be penalized when the Shannon entropy measure is applied. There are a lot of problems with this staring point that I'll get to, but I would assign a symbol to each "function" like a=next line, b=*, e=OR, f=space between arguments, and so on. Numbers would remain number symbols 1,2,3, because they are already efficiently encoded. Then the Shannon entropy (in bits) and therefore the "reduced" level attempted Kolgomorov complexity (Shannon entropy in bits) is
where i=1 to n is for each of n distinct symbols, is "frequency of symbol occurring in program", and is the number of times symbol occurs in the program that is symbols long. This is the Shannon bits (the unit is called "shannons") in the program. The reason this is a better measure of k is because if
is < 1, then there is an inefficiency in the way the symbols themselves were used (which has no relevance to the efficiency of the logic of the algorithm) that is more efficient when expressed as bits. Notice the log base is "n" in this 2nd equation. This H is normalized and is the truest entropy per symbol. To calculate things in this base use logn(x) = ln(x)/ln(n)
But there is a problem. First let me say the research indicates it is only a small problem for any reasonable language that requires only a constant to be added based on the language, which made k-complexity much more interesting and immediately useful. Suppose you want to define a function to be equal to a longer program and just assign a symbol to it, so the program length is reduced to nearly "1". Or maybe a certain function in a certain language is more efficient for certain problems. So there needs to be a reference "language" (set of allowable functions) to compare one K program to the next. All standard functions have a known optimal expression in Boolean logic: AND, NOT, and OR. So by choosing those 3 as the only functions, any program in any higher-level language can be reduced back to a standard. Going even further, these can be reduced back to a transistor count or to a single universal logic gate like NAND or XOR, or a single universal reversible logic gate like Toffoli or Fredkin. Transistors are just performing a function to, so i guess they are universal to, but I am not sure. The universal gates can be wired to performs all logic operations and are Turing complete. So any program in any language can be re-expressed into a single fundamental function. The function itself will not even need a symbol in the program to identify it (as I'll show), so that the only code in the program is describing the wiring between "addresses". So the complexity describes the complexity of the physical connections, the path the electrons or photons in the program take, without regard to the distance between them. Example: sequence of symbols ABCCDA&CEFA&G-A would mean "perform A NAND B and send output to C, perform C NAND D and send output to A and C, perform E NAND F and send output to A and F, go to A. The would define input and output addresses, I would use something like ABD>(program)>G. The "goto" symbol "-" is not used if the program is expanded to express Levine's modified K complexity which penalizes loop \s by expanding them. It thereby takes into account computational time (and therefore a crude measure of energy energy) as well and a different type of program length.
The program length by itself could or should be a measure of the computational infrastructure required (measured as energy required to create the transistors), which is another reason it should be in terms of something like a single gate: so it's cost to implement can be measured. What's the cost to build an OR verses a fourier transform? Answer: Count the required transistors or the NAND gates (which are made up of a distinct number of transistors). All basic functions already have had their Boolean logic and transistors reduce to a minimal level so it's not arbitrary or difficult.
I think this is how you can get a comparable K in all programs in all languages: express them in a single function with distinct symbols representing the input and ouput addresses. Then calculate Shannon's N*H to get its K or Levine complexity in absolute physical terms. Using Landauer's principle, counting the number of times a bit address (at the input or output of a NAND gate) changes state will give the total energy in joules that the program used, Joules=k*T*ln(2)*N where k is larger than Boltzmann's constant kb as a measure of the inefficiency of the particular physical implementation of the NAND gates. If a theoretically reversible gate is used there is theoretically no computational energy loss, except for the input and output changes. Ywaz (talk) 22:20, 12 December 2015 (UTC)[reply]

Inconsistencies in the pseudocode?

[edit]

The pseudocode:

  function GenerateComplexString(int n)
     for i = 1 to infinity:
        for each string s of length exactly i
           if KolmogorovComplexity(s) >= n
              return s
              quit

and:

  function GenerateParadoxicalString()
      return GenerateComplexString(n0)

Seem to use two different languages. In the first case "return" means something like "save this value for the calling function but continue executing like normal", which necessitates the "quit" statement directly after. The second example does not have a "quit" statement and instead works like normal programming languages like C, Pascal, Python, etc.

Suggestion: remove the "quit" statements in all the examples to make the pseudocode function more in line with how "return" is normally (always?) interpreted. --Boxed (talk) 12:28, 20 January 2011 (UTC)[reply]

---

I'd like to remove the pseudocode and just write in English and standard math notation. The pseudocode is distracting and doesn't help the article, IMHO. 06:05, 11 February 2011 (UTC) —Preceding unsigned comment added by 71.141.88.54 (talk)

Okay by me. Personally I don't like pseudocode, especially "structured" pseudocode, because it doesn't lend itself to a rapid translation into a flowchart. What I've done is create a "sandbox page" in my namespace, here: User:Wvbailey/Kolmogorov complexity. There you can experiment to your heart's content before inserting it in the article, plus others can go there too if they want and comment. If you find the need for a more formal algorithm I'd recommend unstructured Basic with line numbers coupled with Knuth's presentation style (see the Euclid's algorithm example at Algorithm). Bill Wvbailey (talk) 16:06, 11 February 2011 (UTC)[reply]

Pi

[edit]

Is pi a good example of a Kolmogorov-simple string? After all, it's an infinite, nonrepeating string that can be defined by a very simple rule. Does this warrant inclusion in the article? 212.68.15.66 (talk) 12:39, 2 February 2011 (UTC)[reply]

We can easily find infinite set of strings, which are infinite, nonrepeating and have short definition, but we certainly can not place them all in the article... ;) --CiaPan (talk) 11:40, 3 February 2011 (UTC)[reply]
True dat, but pi is good ^^ 212.68.15.66 (talk) 07:03, 4 February 2011 (UTC)[reply]
is there some notion of time involved in the program that generates the string? E.g. a program to generate the infinite list of Pi digits can be written but we know (believe?) that it will never finish.
On a related note, the short program that generates the mandalbrot image may take a long time to run. Are two programs (string descriptions) of identical length of equal kolmogorov complexity even if we know (believe?) that the second program takes <K> times longer to run? Perhaps this last question is only answerable if we state assumptions about the underlying abstract machine used to run the program? Funkyj (talk) 20:29, 4 May 2011 (UTC)[reply]
Levin's modification to K-complexity called "Kt" adds the logarithm of the running "time" to its K-value. For the infinite pi case he gives it would have to be based on time per digit generated instead of time as the "t". A multiplication has a higher Kt value than addition. But I would use Landauer's principle to calculate the energy required to store the program in memory plus some constant times the number of state changes required to compute it as a measure of the efficiency (energy-simplicity) of a program for a problem as a measure of its intelligence.Ywaz (talk) 09:44, 21 January 2016 (UTC)[reply]
It's kind of trivial that a program to list all the decimal digits of π will never finish. A program to list all the decimal digits of the real number zero will never finish either. Granted, they're sort of monotonous.... --Trovatore (talk) 21:03, 4 May 2011 (UTC)[reply]
thanks for providing a trivial and uninteresting response to my question. Feel free to make an effort at an interesting comment.
E.g. you might comment on how the string "pi to 10 digits of accuracy" and "pi to 100000000 digits of accuracy" can be represented by the same program with slightly different inputs. Does this mean that the string 3.14 is no more complex than 100 digits of pi?
It takes less information to represent 3 than it takes to represent 100. "11" for three vs "1100100" for one-hundred. The complexity of the representation will increase by log(n) where n is the number of digits of pi you want. I believe the article does mention this, actually. — Preceding unsigned comment added by 67.42.85.224 (talk) 19:06, 17 June 2011 (UTC)[reply]
Also, a simple answer like "the measure of K. complexity does not care how long a program that represents a string takes to run, it only cares bout the length of the program" (or what ever is the true statement) would be nice. Funkyj (talk) 21:19, 5 May 2011 (UTC)[reply]

Chaitin

[edit]

An anonymous editor deleted all mentions of Gregory Chaitin from this article, alleging it to be self-promotion. Is there some evidence of self-promotion in editing this article by Gregory Chaitin? Since long before Wikipedia existed, I've heard the name Chaitin associated with this concept. It seems bizarre to exclude Chaitin from an account of this topic. Michael Hardy (talk) 16:36, 4 February 2011 (UTC)[reply]

I'm in total agreement with you. I've read his popularization MetaMath! The quest for Omega and learned a bunch from it (and pp. 204-206 it has an interesting Bibliography). Also I have an old source ca 1980's written by Martin Davis that mentions Chaitin prominently -- that was where I first encountered his name (if we need the source I'll produce it + I think it's in a talk discussion a few up from this one. But I don't think we will. We can discount this edit as grafitti). Bill Wvbailey (talk) 18:48, 4 February 2011 (UTC)[reply]
Chaitin is given to self-promotion (read any of his popular writings to see this) and the article may have gotten into a state that featured his work too prominently compared to that of others. But eliminating all mention of him is going overboard. 71.141.88.54 (talk) 08:15, 10 February 2011 (UTC)[reply]
I'd like to shorten the stuff about Chaitin's incompleteness theorem and AIT and add a summary/citation of some of Panu Raatikainen's exposition[9] about those topics. I'd also like to get rid of the stuff about Mark Burgin and super-recursiveness, which seems irrelevant to the subject of this article. 71.141.88.54 (talk) 08:29, 10 February 2011 (UTC)[reply]
I have no strong feelings one way or the other, but I worry about self-promotion; we'd have to look at your edits. What I'd suggest is leave the Chaitin material alone but edit out Burgin and add material that you think would benefit the article. But rather than do a major edit as "anonymous", why not reveal yourself first by signing up as an editor? You may notice that Michael Hardey and I aren't hiding behind an IP number. BillWvbailey (talk) 15:21, 10 February 2011 (UTC)[reply]

Maximum minimum description length

[edit]

To explain my edit from: "It can be shown that the Kolmogorov complexity of any string cannot be too much larger than the length of the string itself." to "It can be shown that the Kolmogorov complexity of any string cannot be more than a few bytes larger than the length of the string itself."

The former phrase was vague, and might mean a multiplicative factor rather than an additive factor. Any incompressible string can be wrapped in a print function. The extra length is just the length of the print function, which is a few bytes in any reasonable description language. A description language can be specified that automatically prints any string that starts without a recognized language keyword, so nearly all incompressible strings will be printed without any overhead. The remainder require wrapping with a print function, but on average the overhead will be negligible. Any string that has a significant proportion of description language keywords will not be incompressible. At any rate, a string that is designed to require maximum overhead beyond its own length in a given description language can be represented in another description language with no more than a few bytes to specify the choice of that description language.Enon (talk) 02:49, 26 July 2010 (UTC)[reply]

Hmmm. Depending on quoting characters required, it might actually require multiples. Depending on the description language, it may be possible that the simple description might end up being a multiple. You may be right that in "a natural" programming language, the shortest description would be a bounded quantity longer than the string, but it isn't obvious, and would require a source. — Arthur Rubin (talk) 07:34, 23 September 2011 (UTC)[reply]
It's true in any universal code because it's true in the trivial code $f$ such that $f(s) =s$ for every string $s$. By the definition of universality, for any effective code there is a constant such that the universal system is no worse than that fixed code plus a constant. So for any universal system used to define $K$ there is a constant $C$ such that $K(s) \leq |s| + C$. — Carl (CBM · talk) 11:52, 23 September 2011 (UTC)[reply]

minimum desciption length

[edit]

""namely "ab 32 times", which consists of 11 characters"" according to w.org measuring unit rules should write as ""namely "ab 32 time"s, which consists of 10 characters. ...just an observation 86.121.67.182 (talk) 19:23, 22 September 2011 (UTC)Paul[reply]

Actually, not, if you're rational. And, besides, it should be "'ab' 32 times", making it 13 characters, because of the required quotes. — Arthur Rubin (talk) 07:34, 23 September 2011 (UTC)[reply]
Why not just say "ab×32", which is 5 characters? - dcljr (talk) 20:03, 6 December 2013 (UTC)[reply]
That would be abbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbb. One might accept "'ab'x32" which is 7 characters. Specific representation schemes do not have any place in this article, though. — Arthur Rubin (talk) 02:38, 7 December 2013 (UTC)[reply]

Simpler semi-formal proof of incomputability of Kolmogorov complexity

[edit]

I suggest the following rephrasing of section Kolmogorov complexity#Incomputability of Kolmogorov complexity. My wording may still need improvement, but I'd like to suggest the following changes:

  1. The proof is considerably simpler by using fictitious concrete length values, and I believe the technical subtleties about n growing faster than log(n) in the current version of the proof are not really needed, since the proof is (and should be) somewhat informal, anyway.
  2. The argument for the termination of GenerateComplexString should be stated as an explicit theorem (the 1st one in my version below).
  3. I experimented with a light-blue "word count" in the Berry-paradox sentence, and I think it looks nice.

All these issues are independent of each other.

******************** BEGIN OF SUGGESTED TEXT ********************

Theorem: There exist strings of arbitrary large Kolmogorov complexity. Formally: for each n ∈ ℕ, there is a string s with K(s) ≥ n.[note 1]

Otherwise all possible strings (infinitely many) could be generated by the (finitely many)[note 2] programs with lower complexity.

Theorem: K is not a computable function.

In other words, there is no program which takes a string s as input and produces the integer K(s) as output. The following indirect proof uses a simple Pascal-like language to denote programs; for sake of proof simplicity assume its description (i.e. an interpreter) to have a length of 1'400'000 bits. Assume for contradiction there is a program

  function KolmogorovComplexity(string s)

which takes as input a string s and returns K(s); for sake of proof simplicity, assume its length to be 7'000'000'000 bits.[note 3] Now, consider the following program of length 1'288 bits:

  function GenerateComplexString()
     for i = 1 to infinity:
        for each string s of length exactly i
           if KolmogorovComplexity(s) >= 8000000000
              return s

Using KolmogorovComplexity as a subroutine, the program tries every string, starting with the shortest, until it returns a string with Kolmogorov complexity at least 8'000'000'000 bits,[note 4] i.e. a string that cannot be produced by any program shorter than 8'000'000'000 bits. However, the overall length of the above program that produced s is only 7'001'401'288 bits, which is a contradiction. (If the code of KolmogorovComplexity has any other length, the constant used in GenerateComplexString can always be changed appropriately.)

This proof uses a contradiction similar to that of the Berry paradox: "1The 2smallest 3positive 4integer 5that 6cannot 7be 8defined 9in 10fewer 11than 12twenty 13English 14words". It is also possible to show the non-computability of K by reduction from the non-computability of the halting problem H, since K and H are Turing-equivalent.[1]

In the programming language community there is a corollary known as the full employment theorem, stating that there is no perfect size-optimizing compiler.

References

  1. ^ [1]

Notes

  1. ^ However, an s with K(s) = n needn't exist for every n. For example, if n isn't a multiple of 7 bits, no ASCII program can have length n.
  2. ^ There are 1 + 128 + 1282 + ... + 128n = 128n+1 − 1/127 different 7-bit-ASCII program texts of length up to 7⋅n bits; cf. geometric series.
  3. ^ possibly padded with spaces or nops
  4. ^ By the previous theorem, such a string exists.

******************** END OF SUGGESTED TEXT ********************

Jochen Burghardt (talk) 16:58, 25 January 2014 (UTC)[reply]


As there were no objections, I moved the suggested text above to the article today. I intend to come up with a similar simplification of the proof in section Kolmogorov complexity#Chaitin's incompleteness theorem. - Jochen Burghardt (talk) 16:12, 7 February 2014 (UTC)[reply]
Jochen Burghardt, I'm afraid I've been busy, but I do not like the new proof; it does not seem "simpler". If you could rewrite it so it does not use specific numbers, that would be acceptable to me. Otherwise, I feel the previous proof (with the addition of the theorem that there are strings of arbitrarily large complexity) is better. Furthermore, neither proof falls under WP:CALC, so a reference for (either) proof would be needed. Also, course notes are not generally a reliable source; unless the author is published author in the field, they should not be considered a reliable source. As an aside, the numeric notation you are using is a clear violation of WP:MOSNUM#Delimiting (grouping of digits). — Arthur Rubin (talk) 15:45, 13 February 2014 (UTC)[reply]
Sorry if I was too hasty; if you insist, I'll restore the original proof from the 31 Jan version. My intention was to get rid of the additional function GenerateParadoxicalString (with its "free parameter" n0 not belonging to the programming language), and of the fiddling with lengths constants n, n0, U, C, etc. which I find lengthy and difficult to read. If your reason for disliking the concrete-numbers version is that it doesn't mention that n grows faster than log(n), we could add some explanation after "...can always be changed appropriately".
I wasn't yet aware of WP:MOSNUM, thanks for the hint; I'd suggest to use the ((value)) template, if the numbers are kept.
Concerning a reference (the previous version didn't have one either, so I didn't search before your above message), I could only find:
  • M. Li, P.M.B. Vitányi (1990). Jan van Leeuwen (ed.). Kolmogorov Complexity and its Applications. Handbook of Theoretical Computer Science. Vol. A. Elsevier. pp. 187–254. Here: p.207
  • Ming Li, Paul Vitányi (1997). An Introduction to Kolmogorov Complexity and Its Applications. Graduate texts in computer science. Springer. Here: Section 2.3, Theorem 2.3.2, p.121
Both proofs are almost literally equal to each other, but differ from each wikipedia version. Li+Vitányi attribute their proof to Kolmogorov, cited from the survey of
  • A.K. Zvonkin, L.A. Levin (1970). "The Complexity of Finite Objects and the Development of the Concepts of Information and Randomness by Means of the Theory of Algorithms". Russ. Math. Surveys. 25: 83–124.
They prove more generally that no partial recursive function, defined (i.e. its algorithm terminating) on an infinite set, can coincide with K over its whole domain of definition. The proof is quite short and indirect, but the resemblance to Berry's paradox is only implicit there. Therefore, I'd prefer (any version of) the wikipedia version. - Jochen Burghardt (talk) 10:15, 14 February 2014 (UTC)[reply]
Here is the proof from Li+Vitányi (1990):
Assume for contradition a partial recursive function φ, defined on an infinite set, and coinciding there with K. Using a Theorem from computability theory, select an infinite recursive subset A in the domain of the definition of φ. The function f(m) = min { x: K(x)≥mxA } is total recursive, since K(x)=φ(x) on A, and takes arbitrarily large values. Hence
m
K(f(m)) by definition of f(m)
Kf(f(m)) + cf       by the invariance theorem (which doesn't need Turing completeness for Kf in Li+Vitányi's version)
|m| + cf since Kf(f(m)) = min { |p|: f(p)=f(m) } by Li+Vitányi's definition (p.198)
= log(m) + cf since the number m is considered as a bit string
But this is impossible for arbitrarily large m, and cf independent of m.
I think, φ, f, cf, and m corresponds to KolmogorovComplexity, GenerateComplexString, U+C, and n0, respectively, in the article's 31 Jan version. - Jochen Burghardt (talk) 11:43, 16 February 2014 (UTC)[reply]

Mandelbrot set

[edit]

After some study, it appears that whether a (rational) complex number z is in the Mandelbrot set may be undecidable, so it may not be a good example. Perhaps the Koch snowflake? — Arthur Rubin (talk) 06:48, 11 August 2014 (UTC)[reply]

Huh? The image in question was in fact generated by a (presumably small) computer program, so it can necessarily be encoded by a small program. Probably that program has a threshold on the number of iterations it uses before giving up and assuming that the point in question is in the Mandelbrot set. If you have some more precise mathematical definition of the image in mind than "whatever that Mandelbrot program happened to generate" then you need to start worrying about things like anti-aliasing because just going by whether a mathematical point at the center of a pixel is in or out doesn't produce good images. But if you really did want the point at the center of the pixel, you could encode the image precisely by storing a version of the program that hardcoded a good-enough threshold. (Different crops and zooms would require different thresholds, and the threshold itself might not be easy to compute, but we're only talking about existence of a short code so that's not a problem.) —David Eppstein (talk) 07:00, 11 August 2014 (UTC)[reply]
The image looks complicated, and is computable; however, its relationship to the Mandelbrot set is that of being "close". The formal definition of the Mandelbrot set provides a short description, but it does not necessarily represent something computable. However, if none of the points is on the boundary, then represents the set for sufficiently large N; again, we have no algorithm to determine N; log N could be larger than the number of points. As far as I know, the question of whether a square in the complex plane is contained within the Mandelbrot set is believed undecidable. If we make it clear that we don't know whether the image is exactly the Mandelbrot set, it's true that it's likely that the complexity of the description of the program is less than that of the raw image, but we cannot be sure. I still like the Koch snowflake better. — Arthur Rubin (talk) 07:25, 11 August 2014 (UTC)[reply]

Description for the second string

[edit]

As the second string doesn't use all of the 112 possible characters, couldn't it be compressed by prefixing it with a 112 bit number which indicates the characters that will be used and then have the rest of it transformed into a lower base, reducing its size. If you assume the character set to be only the characters contained in "4c1j5b2p0cv4w1x8rx2y39umgw5q85s7", then it could be spilt in two and then this method could be applied to the two parts separately.


Kolmogorov complexity of Numbers

[edit]

A few attempts have been made to define the Kolmogorov complexity of numbers. https://oeis.org/A168650 defines the complexity of an integer as the length of the shortest expression in C/C++/Java/Python that evaluates to that integer. Also there is a paper about the complexity of real numbers: L. Staiger, "The Kolmogorov complexity of real numbers", Theoretical Computer Science, pp. 455-466, 2002. Should these be added to the article? 118.210.239.185 (talk) 02:00, 29 August 2015 (UTC)[reply]

[edit]

Hello fellow Wikipedians,

I have just added archive links to 3 external links on Kolmogorov complexity. Please take a moment to review my edit. If necessary, add {{cbignore}} after the link to keep me from modifying it. Alternatively, you can add {{nobots|deny=InternetArchiveBot}} to keep me off the page altogether. I made the following changes:

When you have finished reviewing my changes, please set the checked parameter below to true to let others know.

checkY An editor has reviewed this edit and fixed any errors that were found.

  • If you have discovered URLs which were erroneously considered dead by the bot, you can report them with this tool.
  • If you found an error with any archives or the URLs themselves, you can fix them with this tool.

Cheers.—cyberbot IITalk to my owner:Online 07:07, 27 February 2016 (UTC)[reply]

[edit]

Hello fellow Wikipedians,

I have just modified one external link on Kolmogorov complexity. Please take a moment to review my edit. If you have any questions, or need the bot to ignore the links, or the page altogether, please visit this simple FaQ for additional information. I made the following changes:

When you have finished reviewing my changes, you may follow the instructions on the template below to fix any issues with the URLs.

This message was posted before February 2018. After February 2018, "External links modified" talk page sections are no longer generated or monitored by InternetArchiveBot. No special action is required regarding these talk page notices, other than regular verification using the archive tool instructions below. Editors have permission to delete these "External links modified" talk page sections if they want to de-clutter talk pages, but see the RfC before doing mass systematic removals. This message is updated dynamically through the template {{source check}} (last update: 5 June 2024).

  • If you have discovered URLs which were erroneously considered dead by the bot, you can report them with this tool.
  • If you found an error with any archives or the URLs themselves, you can fix them with this tool.

Cheers.—InternetArchiveBot (Report bug) 01:37, 12 December 2017 (UTC)[reply]

"Nontrivial lower bound"

[edit]

See #Incomputable above for context; I think a new section is more convenient to discuss this.

The article currently says:

In particular, for almost all objects, it is not possible to compute even a nontrivial lower bound for its Kolmogorov complexity (see {{Section link}}: required section parameter(s) missing), let alone its exact value.

I'm afraid I still think this text is unclear. Even deferring consideration of what a "nontrivial" lower bound is, or what sense of "almost all" we're using, what does it mean to "compute" a lower bound for a particular object? Or are the quantifiers mixed up here? Should it be something like "any computation that purports to establish a nontrivial lower bound for a given string will answer incorrectly for almost all of them?

One might hope to get these answers from the link (which at least is to the same article, so that's good, as articles must be correct as a self-contained unit), but the link is to the Notes section , and I see nothing there that helps much. Even if it's really supposed to go to the "Chaitin's incompleteness theorem" section, it's kind of obscure how you unpack that to get to the claim as stated. --Trovatore (talk) 20:18, 1 March 2020 (UTC)[reply]

My natural assumption for the meaning of almost all would be: a subset of strings, over whatever alphabet you choose, whose proportion among the strings of length n grows to one in the limit as n goes to infinity. But it could instead easily mean for all but finitely many strings. And my natural assumption for the meaning of "compute" would be to feed it to an algorithm that is guaranteed to terminate with a valid lower bound. "Nontrivial" is less obvious, but it could plausibly mean growing beyond any constant bound. With those assumptions, the formalized version of the statement would be: for every algorithm A that computes on all inputs a valid lower bound on the input's Kolmogorov complexity, there exists a constant b, such that the proportion of strings of length n for which the output of A is at most b goes to one as n goes to infinity. Or alternatively, for every algorithm A ... there exists a constant b, such that A produces an answer greater than b for only finitely many strings. Am I close? However, even if this is accurate, per WP:TECHNICAL, I think it is a lot less intuitive and understandable. To understand it, one basically has to convert it in the other direction to figure out what the intuitive meaning of each part of the statement is. So I don't think we should remove or replace the intuitive statement, but maybe we can also add a little more about how to formulate it mathematically, at least with enough information that someone reading it could recover a formal statement and have more confidence than I do that they've done so accurately. —David Eppstein (talk) 20:35, 1 March 2020 (UTC)[reply]
My general feeling is that, if we can't make it a good deal closer to precise, then we should make it more obviously vague, something like "there are provable limits to any systematic attempt to find lower bounds for Kolmogorov complexity". --Trovatore (talk) 20:54, 1 March 2020 (UTC)[reply]
The link should have pointed to section Kolmogorov complexity#Chaitin's incompleteness theorem, but something went wrong with {{slink}}; I fixed this.
The lead sentence under discussion was intended to be an informal description of the theorem in that section. So "almost all" is meant to denote "all but finitely many", "compute" is meant to denote "obtain in some way by application of rules from a formal system S", and "nontrivial" is meant to exclude lower bounds (like 0) that are "pretty unrelated" to the input string s (I admit that this is rather vague). Another attempt to describe the theorem is this: given a formal system S, some constant L can be determined such that (of course) all but finitely many strings have Kolmogorov complexity ≥L, but (according to the theorem) this isn't provable for any of them; if S includes transitivity of ≥, a provable lower bound, for an arbitrary string whatsoever, must be below L. - To put it in David Eppstein's terminology: for every algorithm A that computes on all inputs a valid lower bound on the input's Kolmogorov complexity, there exists a constant b such that A produces an answer greater than b for no input at all.
Provided my understanding is right, what about the following lead sentence: "In particular, as a consequence of #Chaitin's incompleteness theorem, the Kolmogorov complexity cannot be computed by any algorithm, except possibly for finitely many objects." This avoids the "lower bound", and hence the "nontrivial" issue. On the other hand, I'm not sure that the exact complexity value is uncomputable for all inputs, by some other theorem. - Jochen Burghardt (talk) 23:56, 1 March 2020 (UTC)[reply]
In the Handbook of Th. Comp. Sci. 1990, Li and Vitanyi give (in Sect.2.5, p.207) a theorem by Kolmogorov: "(a) The function K(x) is not partial recursive. Moreover, no partial recursive function f(x), defined on an infinite set of points, can coincide with K(x) over the whole domain of its definition. (b) There is a total recursive function H(t,x), monotonically decreasing in t, such that limt→∞ H(t,x) = K(x). That is, we can obtain arbitrary good estimates for K(x), but not uniformly." - Part (a) seems to be consistent with my above suggestions for a new lead sentence. Apparently, Chaitin's incompleteness theorem is not needed to prove (a), so the lead sentence would describe a proper weakening of Chaitin's incompleteness theorem only. - Jochen Burghardt (talk) 00:10, 2 March 2020 (UTC)[reply]

@Jochen Burghardt:: what is the relevance of Demoscene then? Demoscene is only tangentially related to this topic, while procedural generation sits precisely between this and demoscene. GregorB (talk) 20:04, 25 April 2021 (UTC)[reply]

@GregorB: Demoscene may be relevant since it emphasizes -just like Kolmogorov Complexity- minimal programs (well, at least according to the explanation at Kolmogorov complexity). Procedural generation aparently has no aspect of minimality, or has it?
But the method demoscene is achieving minimality is precisely procedural generation. The rest of what demoscene does is unrelated to the topic. GregorB (talk) 07:33, 26 April 2021 (UTC)[reply]
This doesn't convinve me: By analogy, one could argue "the hardware used by demoscene is a computer, so we should also link computer".
After having read the lead of Demoscene and glanced through the rest, I tent to remove also that link, since there is almost nothing about "smallest programs" in the article (the only exception I found was the first lead sentence). Would it be OK for you to remove both links? - Jochen Burghardt (talk) 09:27, 26 April 2021 (UTC)[reply]
I agree. Kolmogorov complexity is after all a theoretical concept which I believe is connected to procedural generation, but is also independent from it (in the sense that understanding the former does not really help with with the latter). GregorB (talk) 14:06, 27 April 2021 (UTC)[reply]

Big O notation

[edit]

Looking at the recent edit in Kolmogorov_complexity#Chain rule for Kolmogorov complexity, I wonder what O(...) might mean when no functions are involved. (According to Big O notation, it describes the limiting behavior of a function when the argument tends towards a particular value or infinity.) Can anybody explain? - Jochen Burghardt (talk) 17:21, 17 June 2023 (UTC)[reply]

This is in complexity theory, so it's
There exists a constant c such that for all X and Y,
K(X,Y) < K(X) + K(Y|X) + c*max(1,log(K(X,Y))).
JumpDiscont (talk) 00:49, 4 April 2024 (UTC)[reply]

Occam's Razor?

[edit]

KC is essentially a formalization of Occam's Razor -- given that, it seems quite odd that Occam's Razor is not even mentioned in the article until the External Links section all the way at the bottom. It seems like the kind of thing that should be in the opening paragraph, or at least somewhere in the text. Or am I missing sth? --2A02:8071:7141:6400:9DC6:E5A6:8DEB:FF7E (talk) 09:21, 26 May 2024 (UTC)[reply]