Update cookies preferences

E-book: Information-Theoretic Approach to Neural Computing

  • Format - PDF+DRM
  • Price: 110,53 €*
  • * the price is final i.e. no additional discount will apply
  • Add to basket
  • Add to Wishlist
  • This ebook is for personal use only. E-Books are non-refundable.

DRM restrictions

  • Copying (copy/paste):

    not allowed

  • Printing:

    not allowed

  • Usage:

    Digital Rights Management (DRM)
    The publisher has supplied this book in encrypted form, which means that you need to install free software in order to unlock and read it.  To read this e-book you have to create Adobe ID More info here. Ebook can be read and downloaded up to 6 devices (single user with the same Adobe ID).

    Required software
    To read this ebook on a mobile device (phone or tablet) you'll need to install this free app: PocketBook Reader (iOS / Android)

    To download and read this eBook on a PC or Mac you need Adobe Digital Editions (This is a free app specially developed for eBooks. It's not the same as Adobe Reader, which you probably already have on your computer.)

    You can't read this ebook with Amazon Kindle

Neural networks provide a powerful new technology to model and control nonlinear and complex systems. In this book, the authors present a detailed formulation of neural networks from the information-theoretic viewpoint. They show how this perspective provides new insights into the design theory of neural networks. In particular, they show how these methods may be applied to the topics of supervised and unsupervised learning, including feature extraction, linear and nonlinear independent component analysis, and Boltzmann machines.
Readers are assumed to have a basic understanding of neural networks, but all of the relevant concepts from information theory are carefully introduced and explained. Consequently, readers from several different scientific disciplines - notably, cognitive scientists, engineers, physicists, statisticians, and computer scientists - will find this book to be a very valuable contribution to this topic.

A detailed formulation of neural networks from the information-theoretic viewpoint. The authors show how this perspective provides new insights into the design theory of neural networks. In particular they demonstrate how these methods may be applied to the topics of supervised and unsupervised learning, including feature extraction, linear and non-linear independent component analysis, and Boltzmann machines. Readers are assumed to have a basic understanding of neural networks, but all the relevant concepts from information theory are carefully introduced and explained. Consequently, readers from varied scientific disciplines, notably cognitive scientists, engineers, physicists, statisticians, and computer scientists, will find this an extremely valuable introduction to this topic.
1 Introduction.- 2 Preliminaries of Information Theory and Neural
Networks.- 2.1 Elements of Information Theory.- 2.2 Elements of the Theory of
Neural Networks.- I: Unsupervised Learning.- 3 Linear Feature Extraction:
Infomax Principle.- 4 Independent Component Analysis: General Formulation and
Linear Case.- 5 Nonlinear Feature Extraction: Boolean Stochastic Networks.- 6
Nonlinear Feature Extraction: Deterministic Neural Networks.- II: Supervised
Learning.- 7 Supervised Learning and Statistical Estimation.- 8 Statistical
Physics Theory of Supervised Learning and Generalization.- 9 Composite
Networks.- 10 Information Theory Based Regularizing Methods.- References.