Variants of Solomonoff induction

From Machinelearning
Revision as of 04:46, 31 March 2019 by IssaRice (talk | contribs)

This page lists some variants of Solomonoff induction.

For determinism, I think "deterministic" is the same as "Solomonoff prior" and "stochastic" is the same as "universal mixture". Sterkenburg calls deterministic versions a "bottom-up approach" whereas the universal mixture is a "top-down approach" (p. 30).[1] For deterministic variants, the type of universal machine must be specified. With universal mixtures, one must specify two things: the weighting to use, and the class of distributions to consider.

For discrete vs continuous, I think this just means whether the prior we define is over finite strings or over infinite sequences (where we want to know the probability of an infinite sequence starting with a given finite string).

Source Formula Determinism Discrete vs continuous Notes
LessWrong Wiki[2] m(y0)=pP:U(p)=y02(p) where P is the set of self-delimiting programs Deterministic; page doesn't say type of machine, but uses self-delimiting programs and it's discrete, so prefix Turing machine? Discrete because of the U(p)=y0 rather than U(p)=y0*
Scholarpedia discrete universal a priori probability[3] m(x)=p:U(p)=x2(p) where the sum is over halting programs deterministic? prefix Turing machine discrete
Scholarpedia continuous universal a priori probability[3] M(x)=p:U(p)=x*2(p) where the sum is over minimal programs deterministic? Monotone Turing machine Continuous
Sterkenburg (p. 22)[1] PI(σ)=limn|Tσ,n||Tn| where σ is a finite string, Tn is the set of all halting (valid) inputs of length n to the reference machine U, Tσ,n is the set of all halting (valid) inputs of length n that output something starting with σ deterministic; universal Turing machine (no restrictions on prefix-free-ness) discrete?
Sterkenburg (p. 24)[1] P'II(σ)=2|τmin| where τmin is the shortest program τ such that U(τ)=σ (i.e. the shortest program that causes the reference machine to output σ and halt) deterministic; universal Turing machine, universal prefix machine (to get a probability distribution; see remark on p. 27) discrete? this formula does not define a probability distribution over strings σ because the sum of probabilities does not converge
Sterkenburg (p. 25)[1] P'II(σ)=limnτTσ,n2|τ| where Tσ,n is the set of all programs τ of length n such that U(τ) begins with σ deterministic; universal Turing machine P'II(σ) is divergent even for a single σ, so this is not actually a workable version, but is intended as a stepping stone
Sterkenburg (p. 26)[1] PII(σ)=limϵ0limnτTσ,n(1ϵ2)|τ| deterministic; universal Turing machine, universal prefix machine (to get a probability distribution; see remark on p. 27) The use of the ϵ is a hack to get the sum to converge
Sterkenburg (p. 29)[1] QU(σ)=τTσ2|τ| where Tσ is the set of minimal descriptions of σ (i.e. set of programs that output something starting with σ such that if one removes one bit from the end of the program, it no longer outputs something starting with σ) deterministic; universal monotone machine continuous?
Sterkenburg (p. 31)[1] PIV(σ)=limnifi,nPi(σ) Stochastic;
Sterkenburg (p. 33)[1] ξw(σ)=iw(μi)μi(σ) Stochastic; weighting unspecified, except for the requirement that w(μi)>0 for all i and that iw(μi)1; the model class is all semicomputable semimeasures

References

  1. 1.0 1.1 1.2 1.3 1.4 1.5 1.6 1.7 Tom Florian Sterkenburg. "The Foundations of Solomonoff Prediction". February 2013.
  2. https://wiki.lesswrong.com/wiki/Solomonoff_induction
  3. 3.0 3.1 Marcus Hutter; Shane Legg; Paul M.B. Vitanyi. "Algorithmic probability". Scholarpedia. 2007.