Solomonoff's theory of inductive inference is a mathematical proof that if a universe is generated by an algorithm, then observations of that universe, encoded as a dataset, are best predicted by the smallest executable archive of that dataset. This formalization of
Occam's razor
Occam's razor, Ockham's razor, or Ocham's razor ( la, novacula Occami), also known as the principle of parsimony or the law of parsimony ( la, lex parsimoniae), is the problem-solving principle that "entities should not be multiplied beyond neces ...
[JJ McCall. Induction: From Kolmogorov and Solomonoff to De Finetti and Back to Kolmogorov – Metroeconomica, 2004 – Wiley Online Library.][D Stork. Foundations of Occam's razor and parsimony in learning from ricoh.com – NIPS 2001 Workshop, 2001][A.N. Soklakov. Occam's razor as a formal basis for a physical theor]
from arxiv.org
– Foundations of Physics Letters, 2002 – Springer[M Hutter. On the existence and convergence of computable universal prior]
arxiv.org
– Algorithmic Learning Theory, 2003 – Springer for
induction was introduced by
Ray Solomonoff, based on
probability theory and
theoretical computer science.
In essence, Solomonoff's induction derives the
posterior probability of any
computable theory, given a sequence of observed data. This posterior probability is derived from
Bayes rule and some ''universal'' prior, that is, a prior that assigns a positive probability to any computable theory.
Origin
Philosophical
The theory is based in philosophical foundations, and was founded by
Ray Solomonoff around 1960. It is a mathematically formalized combination of
Occam's razor
Occam's razor, Ockham's razor, or Ocham's razor ( la, novacula Occami), also known as the principle of parsimony or the law of parsimony ( la, lex parsimoniae), is the problem-solving principle that "entities should not be multiplied beyond neces ...
and the
Principle of Multiple Explanations
Epicurus (; grc-gre, Ἐπίκουρος ; 341–270 BC) was an ancient Greek philosopher and sage who founded Epicureanism, a highly influential school of philosophy. He was born on the Greek island of Samos to Athenian parents. Influenced ...
.
[Ming Li and Paul Vitanyi, ''An Introduction to Kolmogorov Complexity and Its Applications.'' Springer-Verlag, N.Y., 2008p 339 ff.] All
computable theories which perfectly describe previous observations are used to calculate the probability of the next observation, with more weight put on the shorter computable theories. Marcus Hutter's
universal artificial intelligence builds upon this to calculate the
expected value
In probability theory, the expected value (also called expectation, expectancy, mathematical expectation, mean, average, or first moment) is a generalization of the weighted average. Informally, the expected value is the arithmetic mean of a l ...
of an action.
Principle
Solomonoff's induction has been argued to be the computational formalization of pure
Bayesianism
Bayesian probability is an interpretation of the concept of probability, in which, instead of frequency or propensity of some phenomenon, probability is interpreted as reasonable expectation representing a state of knowledge or as quantification o ...
.
To understand, recall that Bayesianism derives the posterior probability