Abstract
Solomonoff’s uncomputable universal prediction scheme ξ allows to predict the next symbol x k of a sequence x 1 ...x k — 1 for any Turing computable, but otherwise unknown, probabilistic environment μ. This scheme will be generalized to arbitrary environmental classes, which, among others, allows the construction of computable universal prediction schemes ξ. Convergence of ξ to μ in a conditional mean squared sense and with μ probability 1 is proven. It is shown that the average number of prediction errors made by the universal ξ scheme rapidly converges to those made by the best possible informed μ scheme. The schemes, theorems and proofs are given for general finite alphabet, which results in additional complications as compared to the binary case. Several extensions of the presented theory and results are outlined. They include general loss functions and bounds, games of chance, infinite alphabet, partial and delayed prediction, classification, and more active systems.
This work was supported by SNF grant 2000-61847.00 to Jurgen Schmidhuber.
Chapter PDF
References
D. Angluin and C. H. Smith. Inductive inference: Theory and methods. ACM Computing Surveys, 15(3):237–269, 1983.
C. S. Calude et al. Recursively enumerable reals and Chaitin Ω numbers. In 15th Annual Symposium on Theoretical Aspects of Computer Science, volume 1373 of lncs, pages 596–606, Paris France, 1998. Springer.
G. J. Chaitin. Algorithmic information and evolution. in O.T. Solbrig and G. Nicolis, Perspectives on Biological Complexity, IUBS Press, pages 51–60, 1991.
M. Feder, N. Merhav, and M. Gutman. Universal prediction of individual sequences. IEEE Transactions on Information Theory, 38:1258–1270, 1992.
M. Hutter. New error bounds for Solomonoff prediction. Journal of Computer and System Science, in press, 1999. ftp://ftp.idsia.ch/pub/techrep/IDSIA-11-00.ps.gz.
M. Hutter. A theory of universal artificial intelligence based on algorithmic complexity. Technical report, 62 pages, 2000. http://arxiv.org/abs/cs.AI/0004001.
M. Hutter. Optimality of universal Bayesian prediction for general loss and alphabet. Technical Report IDSIA-09-01, Istituto Dalle Molle di Studi sull’Intelligenza Artificiale, Manno(Lugano), Switzerland, 2001.
A. N. Kolmogorov. Three approaches to the quantitative definition of information. Problems of Information and Transmission, 1(1):1–7, 1965.
S. Kullback. Information Theory and Statistics. Wiley, 1959.
L. A. Levin. Universal sequential search problems. Problems of Information Transmission, 9: 265–266, 1973.
L. A. Levin. Randomness conservation inequalities: Information and independence in mathematical theories. Information and Control, 61:15–37, 1984.
M. Li and P. M. B. Vitányi. Inductive reasoning and Kolmogorov complexity. Journal of Computer and System Sciences, 44:343–384, 1992.
M. Li and P. M. B. Vitányi. An introduction to Kolmogorov complexity and its applications. Springer, 2nd edition, 1997.
J. Schmidhuber. Algorithmic theories of everything. Report IDSIA-20-00, quant-ph/0011122, IDSIA, Manno (Lugano), Switzerland, 2000.
R. J. Solomonoff. A formal theory of inductive inference: Part 1 and 2. Inform. Control, 7: 1–22, 224–254, 1964.
R. J. Solomonoff. Complexity-based induction systems: comparisons and convergence theorems. IEEE Trans. Inform. Theory, IT-24:422–432, 1978.
R. J. Solomonoff. The discovery of algorithmic probability. Journal of Computer and System Sciences, 55(1):73–88, 1997.
P. M. B. Vitányi and M. Li. Minimum description length induction, Bayesianism, and Kolmogorov complexity. IEEE Transactions on Information Theory, 46(2):446–464, 2000.
A. K. Zvonkin and L. A. Levin. The complexity of finite objects and the development of the concepts of information and randomness by means of the theory of algorithms. RMS: Russian Mathematical Surveys, 25(6): 83–124, 1970.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2001 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Hutter, M. (2001). Convergence and Error Bounds for Universal Prediction of Nonbinary Sequences. In: De Raedt, L., Flach, P. (eds) Machine Learning: ECML 2001. ECML 2001. Lecture Notes in Computer Science(), vol 2167. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-44795-4_21
Download citation
DOI: https://doi.org/10.1007/3-540-44795-4_21
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-42536-6
Online ISBN: 978-3-540-44795-5
eBook Packages: Springer Book Archive