[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
10.1145/130385.130416acmconferencesArticle/Chapter ViewAbstractPublication PagescoltConference Proceedingsconference-collections
Article
Free access

On learning noisy threshold functions with finite precision weights

Published: 01 July 1992 Publication History

Abstract

We address the issue of the precision required by an N-input threshold element in order to implement a linearly separable mapping. In distinction with previous work we require only the ability to correctly implement the mapping of P randomly chosen training examples, as opposed to the complete boolean mapping. Our results are obtained within the statistical mechanics approach and are thus average case results as opposed to the worst case analyses in the computational learning theory literature. We show that as long as the fraction P/N is finite, then with probability close to 1 as N →∞ a finite number of bits suffice to implement the mapping. This should be compared to the worst case analysis which requires O(N log N) bits. We also calculate the ability of the constrained network to predict novel examples and compare their predictions to those of an unconstrained network. Finally, we address the issue of the performance of the finite-precision network in the face of noisy training examples.

References

[1]
Cover T. Geometrical and statistical properties of systems of linear inequalities with applications in pattern recognition. IEEE Trans. on Elec. Comp. 14:326-334, 1965.
[2]
Edwards S.F. and P.W. Anderson. The theory of spin glasses. J. Phys. F 5:965-974, 1975.
[3]
Fontanari J.F. and R. Meir. Evolving a learning algorithm for the binary perceptron. NET- WORK 2(4):353-359, 1991.
[4]
Gardner E. The space of interactions in neural network models. Y. Phys. A 21:257-270, 1988.
[5]
Gardner E. and B. Derrida. Optimal storage properties of neural networks, J. Phys. A 21 271
[6]
Gutfreund H. and Y. Stein. Capacity of neural networks with discrete synaptic couplings. J. Phys. A 23:2613-2630, 1990.
[7]
GySrgi G. and N. Tishby. Statistical theory of learning a rule. In Neural Networks and Spin Glasses eds. WK Theumann and R KSberle (World Scientific 1989)
[8]
D. Haussler. Decision theoretic generalizations of the PAC model for neural net and other learning applications. University of California Technical Report UCSC-CRL-91-02.
[9]
Haussler D., M. Kearns and R. Schapire. Bounds on the sample complexity of Bayesian learning using information theory and the VC dimension. Proceedings of the fourth workshop on computational learning theory, 1991.
[10]
van Hemmen J.L. and R.G. Palmer. The replica method and a solvable spin glass model. J. Phys. A, 12(4):563-580, 1979.
[11]
Krauth W. and M. Mezard. Storage capacity of memory networks with binary couplings. J. Physique 50:3057-3066, 1989.
[12]
Levin E., N. Tishby and S. Solla. A statistical approach to learning and generalization in neural networks. In R. Rivest, editor, Proc. 2nd Workshop on computational learning the. ory. Morgan Kaufmann, 1989.
[13]
Meir R. &; J.F. Fontanari. Learning from examples in weight constrained neural networks. J. Phys. A, 25: 1149-1168, 1992.
[14]
Meir R. & J.F. Fontanari. On the calculation of learning curves for inconsistent algorithms. Phys. Rev. A 1992.
[15]
Myhill J. and W.H. Kautz. On the size of weights required for linear-input switching functions. IRE Trans. Elec. Comp., EC- 10(2):288-290, 1961.
[16]
Minsky M.L. and S.A. Papert. Perceptrons, MIT Press 1969.
[17]
Martin G.L. and J.A. Pittman. Recognizing hand-printed letters and digits using backpropagation learning. Neural Computation, 3:258-267, 1991.
[18]
Mezard M., G. Purist and M.A. Virasoro. Spin Glass Theory and Beyond, World Scientific, 1987.
[19]
Natarajan B.K. Machine Learning: a Theoretical Approach. Morgan Kaufmann, 1991.
[20]
Opper M. and D. Haussler. Calculation of the learning curve of Bayes optimal classification algorithm. Proceedings of the fourth workshop on computational learning theory, 75-87, 1991.
[21]
Raghavan P. Learning in threshold networks. In Proceedings of the first workshop on computational learning theory, pp. 19-27, 1988.
[22]
Siu K. & J. Bruck. On the power of threshold circuits with small weights. SIAM J. Disc. Math., 4(3):423-435, 1991.
[23]
Seung S., H. Sompolinsky and N. Tishby. Statistical mechanics of learning from examples. Phys. Rev. A, 1992.
[24]
Venkatesh S.S. Directed drift: a new linear threshold algorithm for learning binary weights on-line. Journal of Computer Science and Systems, 1992.

Cited By

View all
  • (1998)Incremental communication for multilayer neural networksIEEE Transactions on Neural Networks10.1109/72.6550319:1(68-82)Online publication date: 1-Jan-1998

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Conferences
COLT '92: Proceedings of the fifth annual workshop on Computational learning theory
July 1992
452 pages
ISBN:089791497X
DOI:10.1145/130385
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 01 July 1992

Permissions

Request permissions for this article.

Check for updates

Qualifiers

  • Article

Conference

COLT92
Sponsor:
COLT92: 5th Annual Workshop on Computational Learning Theory
July 27 - 29, 1992
Pennsylvania, Pittsburgh, USA

Acceptance Rates

Overall Acceptance Rate 35 of 71 submissions, 49%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)32
  • Downloads (Last 6 weeks)8
Reflects downloads up to 15 Jan 2025

Other Metrics

Citations

Cited By

View all
  • (1998)Incremental communication for multilayer neural networksIEEE Transactions on Neural Networks10.1109/72.6550319:1(68-82)Online publication date: 1-Jan-1998

View Options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Login options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media