next up previous
Next: About this document ... Up: NEURAL PREDICTORS FOR DETECTING Previous: ACKNOWLEDGMENTS

Bibliography

1
H. B. Barlow, T. P. Kaushal, and G. J. Mitchison.
Finding minimum entropy codes.
Neural Computation, 1(3):412-423, 1989.

2
G.J. Chaitin.
On the length of programs for computing finite binary sequences: statistical considerations.
Journal of the ACM, 16:145-159, 1969.

3
G.J. Chaitin.
A theory of program size formally identical to information theory.
Journal of the ACM, 22:329-340, 1975.

4
J. Hartmanis.
Generalized Kolmogorov complexity and the structure of feasible computations.
In Proc. 24th IEEE Symposium on Foundations of Computer Science, pages 439-445, 1983.

5
G. Held.
Data Compression.
Wiley and Sons LTD, New York, 1991.

6
S. Hochreiter and J. Schmidhuber.
Flat minima.
Neural Computation, 9(1):1-42, 1997.

7
S. Hochreiter and J. Schmidhuber.
Long short-term memory.
Neural Computation, 9:1681-1726, 1997.

8
A.N. Kolmogorov.
Three approaches to the quantitative definition of information.
Problems of Information Transmission, 1:1-11, 1965.

9
Y. LeCun.
Une procédure d'apprentissage pour réseau à seuil asymétrique.
Proceedings of Cognitiva 85, Paris, pages 599-604, 1985.

10
L. A. Levin.
Universal sequential search problems.
Problems of Information Transmission, 9(3):265-266, 1973.

11
L. A. Levin.
Laws of information (nongrowth) and aspects of the foundation of probability theory.
Problems of Information Transmission, 10(3):206-210, 1974.

12
M. Li and P. M. B. Vitányi.
An introduction to Kolmogorov complexity and its applications.
In J. van Leeuwen, editor, Handbook of Theoretical Computer Science, pages 188-254. Elsevier Science Publishers B.V., 1990.

13
S. Lindstädt.
Comparison of two unsupervised neural network models for redundancy reduction.
In M. C. Mozer, P. Smolensky, D. S. Touretzky, J. L. Elman, and A. S. Weigend, editors, Proc. of the 1993 Connectionist Models Summer School, pages 308-315. Hillsdale, NJ: Erlbaum Associates, 1993.

14
S. Lindstädt.
Comparison of unsupervised neural networks for redundancy reduction, 1993.
Master's thesis, Dept. of Comp. Sci., University of Colorado at Boulder.

15
M. C. Mozer.
A focused back-propagation algorithm for temporal sequence recognition.
Complex Systems, 3:349-381, 1989.

16
D. B. Parker.
Learning-logic.
Technical Report TR-47, Center for Comp. Research in Economics and Management Sci., MIT, 1985.

17
B. A. Pearlmutter.
Learning state space trajectories in recurrent neural networks.
Neural Computation, 1(2):263-269, 1989.

18
A. J. Robinson and F. Fallside.
The utility driven dynamic error propagation network.
Technical Report CUED/F-INFENG/TR.1, Cambridge University Engineering Department, 1987.

19
D. E. Rumelhart, G. E. Hinton, and R. J. Williams.
Learning internal representations by error propagation.
In Parallel Distributed Processing, volume 1, pages 318-362. MIT Press, 1986.

20
J. Schmidhuber.
A fixed size storage $O(n^3)$ time complexity learning algorithm for fully recurrent continually running networks.
Neural Computation, 4(2):243-248, 1992.

21
J. Schmidhuber.
Learning complex, extended sequences using the principle of history compression.
Neural Computation, 4(2):234-242, 1992.

22
J. Schmidhuber.
Learning factorial codes by predictability minimization.
Neural Computation, 4(6):863-879, 1992.

23
J. Schmidhuber.
Learning unambiguous reduced sequence descriptions.
In J. E. Moody, S. J. Hanson, and R. P. Lippman, editors, Advances in Neural Information Processing Systems 4, pages 291-298. San Mateo, CA: Morgan Kaufmann, 1992.

24
J. Schmidhuber.
Netzwerkarchitekturen, Zielfunktionen und Kettenregel. Habilitationsschrift, Institut für Informatik, Technische Universität München, 1993.

25
J. Schmidhuber.
Discovering solutions with low Kolmogorov complexity and high generalization capability.
In A. Prieditis and S. Russell, editors, Machine Learning: Proceedings of the Twelfth International Conference, pages 488-496. Morgan Kaufmann Publishers, San Francisco, CA, 1995.

26
J. Schmidhuber.
Discovering neural nets with low Kolmogorov complexity and high generalization capability.
Neural Networks, 10(5):857-873, 1997.

27
J. Schmidhuber and S. Heil.
Predictive coding with neural nets: Application to text compression.
In G. Tesauro, D. S. Touretzky, and T. K. Leen, editors, Advances in Neural Information Processing Systems 7, pages 1047 - 1054. MIT Press, Cambridge MA, 1995.

28
J. Schmidhuber and S. Heil.
Sequential neural text compression.
IEEE Transactions on Neural Networks, 7(1):142-146, 1996.

29
J. Schmidhuber, M. C. Mozer, and D. Prelinger.
Continuous history compression.
In H. Hüning, S. Neuhauser, M. Raus, and W. Ritschel, editors, Proc. of Intl. Workshop on Neural Networks, RWTH Aachen, pages 87-95. Augustinus, 1993.

30
J. Schmidhuber and D. Prelinger.
Discovering predictable classifications.
Neural Computation, 5(4):625-635, 1993.

31
J. Schmidhuber, J. Zhao, and N. Schraudolph.
Reinforcement learning with self-modifying policies.
In S. Thrun and L. Pratt, editors, Learning to learn, pages 293-309. Kluwer, 1997.

32
J. Schmidhuber, J. Zhao, and M. Wiering.
Shifting inductive bias with success-story algorithm, adaptive Levin search, and incremental self-improvement.
Machine Learning, 28:105-130, 1997.

33
C. E. Shannon.
A mathematical theory of communication (parts I and II).
Bell System Technical Journal, XXVII:379-423, 1948.

34
R.J. Solomonoff.
A formal theory of inductive inference. Part I.
Information and Control, 7:1-22, 1964.

35
O. Watanabe.
Kolmogorov complexity and computational complexity.
EATCS Monographs on Theoretical Computer Science, Springer, 1992.

36
P. J. Werbos.
Beyond Regression: New Tools for Prediction and Analysis in the Behavioral Sciences.
PhD thesis, Harvard University, 1974.

37
P. J. Werbos.
Generalization of backpropagation with application to a recurrent gas market model.
Neural Networks, 1, 1988.

38
R. J. Williams.
Complexity of exact gradient computation algorithms for recurrent neural networks.
Technical Report Technical Report NU-CCS-89-27, Boston: Northeastern University, College of Computer Science, 1989.

39
R. J. Williams and J. Peng.
An efficient gradient-based algorithm for online training of recurrent network trajectories.
Neural Computation, 4:491-501, 1990.

40
R. J. Williams and D. Zipser.
Experimental analysis of the real-time recurrent learning algorithm.
Connection Science, 1(1):87-111, 1989.

41
I. H. Witten, R. M. Neal, and J. G. Cleary.
Arithmetic coding for data compression.
Communications of the ACM, 30(6):520-540, 1987.

42
A. Wyner and J. Ziv.
Fixed data base version of the Lempel-Ziv data compression algorithm.
IEEE Transactions Information Theory, 37:878-880, 1991.

43
J. Ziv and A. Lempel.
A universal algorithm for sequential data compression.
IEEE Transactions on Information Theory, IT-23(5):337-343, 1977.



Juergen Schmidhuber 2003-02-19