Muutke küpsiste eelistusi

Information-Theoretic Approach to Neural Computing 1st ed. 1996. Corr. 2nd printing 1997 [Kõva köide]

  • Formaat: Hardback, 262 pages, kõrgus x laius: 235x155 mm, kaal: 1270 g, XIV, 262 p., 1 Hardback
  • Sari: Perspectives in Neural Computing
  • Ilmumisaeg: 08-Feb-1996
  • Kirjastus: Springer-Verlag New York Inc.
  • ISBN-10: 0387946667
  • ISBN-13: 9780387946665
  • Kõva köide
  • Hind: 95,19 €*
  • * hind on lõplik, st. muud allahindlused enam ei rakendu
  • Tavahind: 111,99 €
  • Säästad 15%
  • Raamatu kohalejõudmiseks kirjastusest kulub orienteeruvalt 3-4 nädalat
  • Kogus:
  • Lisa ostukorvi
  • Tasuta tarne
  • Tellimisaeg 2-4 nädalat
  • Lisa soovinimekirja
  • Formaat: Hardback, 262 pages, kõrgus x laius: 235x155 mm, kaal: 1270 g, XIV, 262 p., 1 Hardback
  • Sari: Perspectives in Neural Computing
  • Ilmumisaeg: 08-Feb-1996
  • Kirjastus: Springer-Verlag New York Inc.
  • ISBN-10: 0387946667
  • ISBN-13: 9780387946665
Neural networks provide a powerful new technology to model and control nonlinear and complex systems. In this book, the authors present a detailed formulation of neural networks from the information-theoretic viewpoint. They show how this perspective provides new insights into the design theory of neural networks. In particular, they show how these methods may be applied to the topics of supervised and unsupervised learning, including feature extraction, linear and nonlinear independent component analysis, and Boltzmann machines.
Readers are assumed to have a basic understanding of neural networks, but all of the relevant concepts from information theory are carefully introduced and explained. Consequently, readers from several different scientific disciplines - notably, cognitive scientists, engineers, physicists, statisticians, and computer scientists - will find this book to be a very valuable contribution to this topic.

A detailed formulation of neural networks from the information-theoretic viewpoint. The authors show how this perspective provides new insights into the design theory of neural networks. In particular they demonstrate how these methods may be applied to the topics of supervised and unsupervised learning, including feature extraction, linear and non-linear independent component analysis, and Boltzmann machines. Readers are assumed to have a basic understanding of neural networks, but all the relevant concepts from information theory are carefully introduced and explained. Consequently, readers from varied scientific disciplines, notably cognitive scientists, engineers, physicists, statisticians, and computer scientists, will find this an extremely valuable introduction to this topic.
1 Introduction.- 2 Preliminaries of Information Theory and Neural
Networks.- 2.1 Elements of Information Theory.- 2.2 Elements of the Theory of
Neural Networks.- I: Unsupervised Learning.- 3 Linear Feature Extraction:
Infomax Principle.- 4 Independent Component Analysis: General Formulation and
Linear Case.- 5 Nonlinear Feature Extraction: Boolean Stochastic Networks.- 6
Nonlinear Feature Extraction: Deterministic Neural Networks.- II: Supervised
Learning.- 7 Supervised Learning and Statistical Estimation.- 8 Statistical
Physics Theory of Supervised Learning and Generalization.- 9 Composite
Networks.- 10 Information Theory Based Regularizing Methods.- References.