next up previous
Next: About this document ... Up: The New AI: General Previous: Acknowledgments

Bibliography

1
M. Beeson.
Foundations of Constructive Mathematics.
Springer-Verlag, Heidelberg, 1985.

2
J. S. Bell.
On the problem of hidden variables in quantum mechanics.
Rev. Mod. Phys., 38:447-452, 1966.

3
C. H. Bennett and D. P. DiVicenzo.
Quantum information and computation.
Nature, 404(6775):256-259, 2000.

4
C. M. Bishop.
Neural networks for pattern recognition.
Oxford University Press, 1995.

5
L. E. J. Brouwer.
Over de Grondslagen der Wiskunde. Dissertation, Doctoral Thesis, University of Amsterdam, 1907.

6
F. Cajori.
History of mathematics (2nd edition).
Macmillan, New York, 1919.

7
G. Cantor.
Über eine Eigenschaft des Inbegriffes aller reellen algebraischen Zahlen.
Crelle's Journal für Mathematik, 77:258-263, 1874.

8
G.J. Chaitin.
On the length of programs for computing finite binary sequences: statistical considerations.
Journal of the ACM, 16:145-159, 1969.

9
G.J. Chaitin.
A theory of program size formally identical to information theory.
Journal of the ACM, 22:329-340, 1975.

10
G.J. Chaitin.
Algorithmic Information Theory.
Cambridge University Press, Cambridge, 1987.

11
D. Deutsch.
The Fabric of Reality.
Allen Lane, New York, NY, 1997.

12
T. Erber and S. Putterman.
Randomness in quantum mechanics - nature's ultimate cryptogram?
Nature, 318(7):41-43, 1985.

13
H. Everett III.
`Relative State' formulation of quantum mechanics.
Reviews of Modern Physics, 29:454-462, 1957.

14
E. F. Fredkin and T. Toffoli.
Conservative logic.
International Journal of Theoretical Physics, 21(3/4):219-253, 1982.

15
R. V. Freyvald.
Functions and functionals computable in the limit.
Transactions of Latvijas Vlasts Univ. Zinatn. Raksti, 210:6-19, 1977.

16
P. Gács.
On the relation between descriptional complexity and algorithmic probability.
Theoretical Computer Science, 22:71-93, 1983.

17
E. M. Gold.
Limiting recursion.
Journal of Symbolic Logic, 30(1):28-46, 1965.

18
M.B. Green, J.H. Schwarz, and E. Witten.
Superstring Theory.
Cambridge University Press, 1987.

19
S. Hochreiter, A. S. Younger, and P. R. Conwell.
Learning to learn using gradient descent.
In Lecture Notes on Comp. Sci. 2130, Proc. Intl. Conf. on Artificial Neural Networks (ICANN-2001), pages 87-94. Springer: Berlin, Heidelberg, 2001.

20
M. Hutter.
Convergence and error bounds of universal prediction for general alphabet.
Proceedings of the 12th European Conference on Machine Learning (ECML-2001), (TR IDSIA-07-01, cs.AI/0103015), 2001.

21
M. Hutter.
General loss bounds for universal sequence prediction.
In C. E. Brodley and A. P. Danyluk, editors, Proceedings of the $18^{th}$ International Conference on Machine Learning (ICML-2001), pages 210-217. Morgan Kaufmann, 2001.
TR IDSIA-03-01, IDSIA, Switzerland, Jan 2001, cs.AI/0101019.

22
M. Hutter.
Towards a universal theory of artificial intelligence based on algorithmic probability and sequential decisions.
Proceedings of the 12$^{th}$ European Conference on Machine Learning (ECML-2001), (TR IDSIA-14-00, cs.AI/0012011):226-238, 2001.

23
M. Hutter.
The fastest and shortest algorithm for all well-defined problems.
International Journal of Foundations of Computer Science, 13(3):431-443, 2002.

24
M. Hutter.
Self-optimizing and Pareto-optimal policies in general environments based on Bayes-mixtures.
In J. Kivinen and R. H. Sloan, editors, Proceedings of the 15th Annual Conference on Computational Learning Theory (COLT 2002), Lecture Notes in Artificial Intelligence, pages 364-379, Sydney, Australia, 2002. Springer.

25
M. Hutter.
A gentle introduction to the universal algorithmic agent AIXI.
In B. Goertzel and C. Pennachin, editors, Real AI: New Approaches to Artificial General Intelligence. Plenum Press, New York, 2003.
To appear.

26
M. I. Jordan and D. E. Rumelhart.
Supervised learning with a distal teacher.
Technical Report Occasional Paper #40, Center for Cog. Sci., Massachusetts Institute of Technology, 1990.

27
L.P. Kaelbling, M.L. Littman, and A.W. Moore.
Reinforcement learning: a survey.
Journal of AI research, 4:237-285, 1996.

28
A.N. Kolmogorov.
Three approaches to the quantitative definition of information.
Problems of Information Transmission, 1:1-11, 1965.

29
L. A. Levin.
Universal sequential search problems.
Problems of Information Transmission, 9(3):265-266, 1973.

30
L. A. Levin.
Laws of information (nongrowth) and aspects of the foundation of probability theory.
Problems of Information Transmission, 10(3):206-210, 1974.

31
M. Li and P. M. B. Vitányi.
An Introduction to Kolmogorov Complexity and its Applications (2nd edition).
Springer, 1997.

32
L. Löwenheim.
Über Möglichkeiten im Relativkalkül.
Mathematische Annalen, 76:447-470, 1915.

33
N. Merhav and M. Feder.
Universal prediction.
IEEE Transactions on Information Theory, 44(6):2124-2147, 1998.

34
T. Mitchell.
Machine Learning.
McGraw Hill, 1997.

35
C. H. Moore and G. C. Leach.
FORTH - a language for interactive computing, 1970.
http://www.ultratechnology.com.

36
A. Newell and H. Simon.
GPS, a program that simulates human thought.
In E. Feigenbaum and J. Feldman, editors, Computers and Thought, pages 279-293. McGraw-Hill, New York, 1963.

37
Nguyen and B. Widrow.
The truck backer-upper: An example of self learning in neural networks.
In Proceedings of the International Joint Conference on Neural Networks, pages 357-363. IEEE Press, 1989.

38
R. Penrose.
The Emperor's New Mind.
Oxford University Press, 1989.

39
K. R. Popper.
The Logic of Scientific Discovery.
Hutchinson, London, 1934.

40
H. Putnam.
Trial and error predicates and the solution to a problem of Mostowski.
Journal of Symbolic Logic, 30(1):49-57, 1965.

41
J. Rissanen.
Stochastic complexity and modeling.
The Annals of Statistics, 14(3):1080-1100, 1986.

42
H. Rogers, Jr.
Theory of Recursive Functions and Effective Computability.
McGraw-Hill, New York, 1967.

43
P. S. Rosenbloom, J. E. Laird, and A. Newell.
The SOAR Papers.
MIT Press, 1993.

44
D. E. Rumelhart, G. E. Hinton, and R. J. Williams.
Learning internal representations by error propagation.
In D. E. Rumelhart and J. L. McClelland, editors, Parallel Distributed Processing, volume 1, pages 318-362. MIT Press, 1986.

45
C. Schmidhuber.
Strings from logic.
Technical Report CERN-TH/2000-316, CERN, Theory Division, 2000.
http://xxx.lanl.gov/abs/hep-th/0011065.

46
J. Schmidhuber.
Reinforcement learning in Markovian and non-Markovian environments.
In D. S. Lippman, J. E. Moody, and D. S. Touretzky, editors, Advances in Neural Information Processing Systems 3, pages 500-506. Morgan Kaufmann, 1991.

47
J. Schmidhuber.
Discovering solutions with low Kolmogorov complexity and high generalization capability.
In A. Prieditis and S. Russell, editors, Machine Learning: Proceedings of the Twelfth International Conference, pages 488-496. Morgan Kaufmann Publishers, San Francisco, CA, 1995.

48
J. Schmidhuber.
A computer scientist's view of life, the universe, and everything.
In C. Freksa, M. Jantzen, and R. Valk, editors, Foundations of Computer Science: Potential - Theory - Cognition, volume 1337, pages 201-208. Lecture Notes in Computer Science, Springer, Berlin, 1997.
Submitted 1996.

49
J. Schmidhuber.
Discovering neural nets with low Kolmogorov complexity and high generalization capability.
Neural Networks, 10(5):857-873, 1997.

50
J. Schmidhuber.
Algorithmic theories of everything.
Technical Report IDSIA-20-00, quant-ph/0011122, IDSIA, Manno (Lugano), Switzerland, 2000.

51
J. Schmidhuber.
Sequential decision making based on direct search.
In R. Sun and C. L. Giles, editors, Sequence Learning: Paradigms, Algorithms, and Applications. Springer, 2001.
Lecture Notes on AI 1828.

52
J. Schmidhuber.
Bias-optimal incremental problem solving.
In Advances in Neural Information Processing Systems 15. MIT Press, Cambridge, MA, 2002.
To appear.

53
J. Schmidhuber.
Hierarchies of generalized Kolmogorov complexities and nonenumerable universal measures computable in the limit.
International Journal of Foundations of Computer Science, 13(4):587-612, 2002.

54
J. Schmidhuber.
Optimal ordered problem solver.
Technical Report IDSIA-12-02, arXiv:cs.AI/0207097 v1, IDSIA, Manno-Lugano, Switzerland, July 2002.

55
J. Schmidhuber.
The Speed Prior: a new simplicity measure yielding near-optimal computable predictions.
In J. Kivinen and R. H. Sloan, editors, Proceedings of the 15th Annual Conference on Computational Learning Theory (COLT 2002), Lecture Notes in Artificial Intelligence, pages 216-228. Springer, Sydney, Australia, 2002.

56
J. Schmidhuber and M. Hutter.
NIPS 2002 workshop on universal learning algorithms and optimal search. Additional speakers: R. Solomonoff, P. M. B. Vitányi, N. Cesa-Bianchi, I. Nemenmann. Whistler, CA, 2002.

57
J. Schmidhuber, J. Zhao, and M. Wiering.
Shifting inductive bias with success-story algorithm, adaptive Levin search, and incremental self-improvement.
Machine Learning, 28:105-130, 1997.

58
T. Skolem.
Logisch-kombinatorische Untersuchungen über Erfüllbarkeit oder Beweisbarkeit mathematischer Sätze nebst einem Theorem über dichte Mengen.
Skrifter utgit av Videnskapsselskapet in Kristiania, I, Mat.-Nat. Kl., N4:1-36, 1919.

59
R.J. Solomonoff.
A formal theory of inductive inference. Part I.
Information and Control, 7:1-22, 1964.

60
R.J. Solomonoff.
Complexity-based induction systems.
IEEE Transactions on Information Theory, IT-24(5):422-432, 1978.

61
R.J. Solomonoff.
An application of algorithmic probability to problems in artificial intelligence.
In L. N. Kanal and J. F. Lemmer, editors, Uncertainty in Artificial Intelligence, pages 473-491. Elsevier Science Publishers, 1986.

62
R.J. Solomonoff.
A system for incremental learning based on algorithmic probability.
In Proceedings of the Sixth Israeli Conference on Artificial Intelligence, Computer Vision and Pattern Recognition, pages 515-527. Tel Aviv, Israel, 1989.

63
G. 't Hooft.
Quantum gravity as a dissipative deterministic system.
Technical Report SPIN-1999/07/gr-gc/9903084, http://xxx.lanl.gov/abs/gr-qc/9903084, Institute for Theoretical Physics, Univ. of Utrecht, and Spinoza Institute, Netherlands, 1999.
Also published in Classical and Quantum Gravity 16, 3263.

64
A. M. Turing.
On computable numbers, with an application to the Entscheidungsproblem.
Proceedings of the London Mathematical Society, Series 2, 41:230-267, 1936.

65
S. Ulam.
Random processes and transformations.
In Proceedings of the International Congress on Mathematics, volume 2, pages 264-275, 1950.

66
V. Vapnik.
The Nature of Statistical Learning Theory.
Springer, New York, 1995.

67
J. von Neumann.
Theory of Self-Reproducing Automata.
University of Illionois Press, Champain, IL, 1966.

68
C. S. Wallace and D. M. Boulton.
An information theoretic measure for classification.
Computer Journal, 11(2):185-194, 1968.

69
P. J. Werbos.
Beyond Regression: New Tools for Prediction and Analysis in the Behavioral Sciences.
PhD thesis, Harvard University, 1974.

70
P. J. Werbos.
Learning how the world works: Specifications for predictive networks in robots and brains.
In Proceedings of IEEE International Conference on Systems, Man and Cybernetics, N.Y., 1987.

71
M.A. Wiering and J. Schmidhuber.
Solving POMDPs with Levin search and EIRA.
In L. Saitta, editor, Machine Learning: Proceedings of the Thirteenth International Conference, pages 534-542. Morgan Kaufmann Publishers, San Francisco, CA, 1996.

72
S. Wolfram.
Universality and complexity in cellular automata.
Physica D, 10:1-35, 1984.

73
K. Zuse.
Rechnender Raum.
Elektronische Datenverarbeitung, 8:336-344, 1967.

74
K. Zuse.
Rechnender Raum.
Friedrich Vieweg & Sohn, Braunschweig, 1969.

75
A. K. Zvonkin and L. A. Levin.
The complexity of finite objects and the algorithmic concepts of information and randomness.
Russian Math. Surveys, 25(6):83-124, 1970.



Juergen Schmidhuber 2003-02-04