Next: About this document ...
Up: NEURAL PREDICTORS FOR DETECTING
Previous: ACKNOWLEDGMENTS
-
- 1
-
H. B. Barlow, T. P. Kaushal, and G. J. Mitchison.
Finding minimum entropy codes.
Neural Computation, 1(3):412-423, 1989.
- 2
-
G.J. Chaitin.
On the length of programs for computing finite binary sequences:
statistical considerations.
Journal of the ACM, 16:145-159, 1969.
- 3
-
G.J. Chaitin.
A theory of program size formally identical to information theory.
Journal of the ACM, 22:329-340, 1975.
- 4
-
J. Hartmanis.
Generalized Kolmogorov complexity and the structure of feasible
computations.
In Proc. 24th IEEE Symposium on Foundations of Computer
Science, pages 439-445, 1983.
- 5
-
G. Held.
Data Compression.
Wiley and Sons LTD, New York, 1991.
- 6
-
S. Hochreiter and J. Schmidhuber.
Flat minima.
Neural Computation, 9(1):1-42, 1997.
- 7
-
S. Hochreiter and J. Schmidhuber.
Long short-term memory.
Neural Computation, 9:1681-1726, 1997.
- 8
-
A.N. Kolmogorov.
Three approaches to the quantitative definition of information.
Problems of Information Transmission, 1:1-11, 1965.
- 9
-
Y. LeCun.
Une procédure d'apprentissage pour réseau à seuil
asymétrique.
Proceedings of Cognitiva 85, Paris, pages 599-604, 1985.
- 10
-
L. A. Levin.
Universal sequential search problems.
Problems of Information Transmission, 9(3):265-266, 1973.
- 11
-
L. A. Levin.
Laws of information (nongrowth) and aspects of the foundation of
probability theory.
Problems of Information Transmission, 10(3):206-210, 1974.
- 12
-
M. Li and P. M. B. Vitányi.
An introduction to Kolmogorov complexity and its applications.
In J. van Leeuwen, editor, Handbook of Theoretical Computer
Science, pages 188-254. Elsevier Science Publishers B.V., 1990.
- 13
-
S. Lindstädt.
Comparison of two unsupervised neural network models for redundancy
reduction.
In M. C. Mozer, P. Smolensky, D. S. Touretzky, J. L. Elman, and A. S.
Weigend, editors, Proc. of the 1993 Connectionist Models Summer School,
pages 308-315. Hillsdale, NJ: Erlbaum Associates, 1993.
- 14
-
S. Lindstädt.
Comparison of unsupervised neural networks for redundancy reduction,
1993.
Master's thesis, Dept. of Comp. Sci., University of Colorado at
Boulder.
- 15
-
M. C. Mozer.
A focused back-propagation algorithm for temporal sequence
recognition.
Complex Systems, 3:349-381, 1989.
- 16
-
D. B. Parker.
Learning-logic.
Technical Report TR-47, Center for Comp. Research in Economics and
Management Sci., MIT, 1985.
- 17
-
B. A. Pearlmutter.
Learning state space trajectories in recurrent neural networks.
Neural Computation, 1(2):263-269, 1989.
- 18
-
A. J. Robinson and F. Fallside.
The utility driven dynamic error propagation network.
Technical Report CUED/F-INFENG/TR.1, Cambridge University Engineering
Department, 1987.
- 19
-
D. E. Rumelhart, G. E. Hinton, and R. J. Williams.
Learning internal representations by error propagation.
In Parallel Distributed Processing, volume 1, pages 318-362.
MIT Press, 1986.
- 20
-
J. Schmidhuber.
A fixed size storage time complexity learning algorithm
for fully recurrent continually running networks.
Neural Computation, 4(2):243-248, 1992.
- 21
-
J. Schmidhuber.
Learning complex, extended sequences using the principle of history
compression.
Neural Computation, 4(2):234-242, 1992.
- 22
-
J. Schmidhuber.
Learning factorial codes by predictability minimization.
Neural Computation, 4(6):863-879, 1992.
- 23
-
J. Schmidhuber.
Learning unambiguous reduced sequence descriptions.
In J. E. Moody, S. J. Hanson, and R. P. Lippman, editors, Advances in Neural Information Processing Systems 4, pages 291-298. San
Mateo, CA: Morgan Kaufmann, 1992.
- 24
-
J. Schmidhuber.
Netzwerkarchitekturen, Zielfunktionen und Kettenregel.
Habilitationsschrift, Institut für Informatik, Technische
Universität München, 1993.
- 25
-
J. Schmidhuber.
Discovering solutions with low Kolmogorov complexity and high
generalization capability.
In A. Prieditis and S. Russell, editors, Machine Learning:
Proceedings of the Twelfth International Conference, pages 488-496. Morgan
Kaufmann Publishers, San Francisco, CA, 1995.
- 26
-
J. Schmidhuber.
Discovering neural nets with low Kolmogorov complexity and high
generalization capability.
Neural Networks, 10(5):857-873, 1997.
- 27
-
J. Schmidhuber and S. Heil.
Predictive coding with neural nets: Application to text compression.
In G. Tesauro, D. S. Touretzky, and T. K. Leen, editors, Advances in Neural Information Processing Systems 7, pages 1047 - 1054. MIT
Press, Cambridge MA, 1995.
- 28
-
J. Schmidhuber and S. Heil.
Sequential neural text compression.
IEEE Transactions on Neural Networks, 7(1):142-146, 1996.
- 29
-
J. Schmidhuber, M. C. Mozer, and D. Prelinger.
Continuous history compression.
In H. Hüning, S. Neuhauser, M. Raus, and W. Ritschel, editors,
Proc. of Intl. Workshop on Neural Networks, RWTH Aachen, pages 87-95.
Augustinus, 1993.
- 30
-
J. Schmidhuber and D. Prelinger.
Discovering predictable classifications.
Neural Computation, 5(4):625-635, 1993.
- 31
-
J. Schmidhuber, J. Zhao, and N. Schraudolph.
Reinforcement learning with self-modifying policies.
In S. Thrun and L. Pratt, editors, Learning to learn, pages
293-309. Kluwer, 1997.
- 32
-
J. Schmidhuber, J. Zhao, and M. Wiering.
Shifting inductive bias with success-story algorithm, adaptive
Levin search, and incremental self-improvement.
Machine Learning, 28:105-130, 1997.
- 33
-
C. E. Shannon.
A mathematical theory of communication (parts I and II).
Bell System Technical Journal, XXVII:379-423, 1948.
- 34
-
R.J. Solomonoff.
A formal theory of inductive inference. Part I.
Information and Control, 7:1-22, 1964.
- 35
-
O. Watanabe.
Kolmogorov complexity and computational complexity.
EATCS Monographs on Theoretical Computer Science, Springer, 1992.
- 36
-
P. J. Werbos.
Beyond Regression: New Tools for Prediction and Analysis in the
Behavioral Sciences.
PhD thesis, Harvard University, 1974.
- 37
-
P. J. Werbos.
Generalization of backpropagation with application to a recurrent gas
market model.
Neural Networks, 1, 1988.
- 38
-
R. J. Williams.
Complexity of exact gradient computation algorithms for recurrent
neural networks.
Technical Report Technical Report NU-CCS-89-27, Boston: Northeastern
University, College of Computer Science, 1989.
- 39
-
R. J. Williams and J. Peng.
An efficient gradient-based algorithm for online training of
recurrent network trajectories.
Neural Computation, 4:491-501, 1990.
- 40
-
R. J. Williams and D. Zipser.
Experimental analysis of the real-time recurrent learning algorithm.
Connection Science, 1(1):87-111, 1989.
- 41
-
I. H. Witten, R. M. Neal, and J. G. Cleary.
Arithmetic coding for data compression.
Communications of the ACM, 30(6):520-540, 1987.
- 42
-
A. Wyner and J. Ziv.
Fixed data base version of the Lempel-Ziv data compression
algorithm.
IEEE Transactions Information Theory, 37:878-880, 1991.
- 43
-
J. Ziv and A. Lempel.
A universal algorithm for sequential data compression.
IEEE Transactions on Information Theory, IT-23(5):337-343,
1977.
Juergen Schmidhuber
2003-02-19