283,99 €
inkl. MwSt.
Versandkostenfrei*
Versandfertig in über 4 Wochen
  • Gebundenes Buch

Durch die Anwendung rückbezüglicher neuronaler Netze läßt sich die Leistungsfähigkeit konventioneller Technologien der digitalen Datenverarbeitung signifikant erhöhen. Von besonderer Bedeutung ist dies für komplexe Aufgaben, wie z.B. die mobile Kommunikation, die Robotik und die Medizintechnik. Das Buch faßt Originalarbeiten zur Stabilität neuronaler Netze zusammen und verbindet streng mathematische Analysen mit anschaulichen Anwendungen und experimentellen Belegen.
New technologies in engineering, physics and biomedicine are demanding increasingly complex methods of digital signal
…mehr

Produktbeschreibung
Durch die Anwendung rückbezüglicher neuronaler Netze läßt sich die Leistungsfähigkeit konventioneller Technologien der digitalen Datenverarbeitung signifikant erhöhen. Von besonderer Bedeutung ist dies für komplexe Aufgaben, wie z.B. die mobile Kommunikation, die Robotik und die Medizintechnik. Das Buch faßt Originalarbeiten zur Stabilität neuronaler Netze zusammen und verbindet streng mathematische Analysen mit anschaulichen Anwendungen und experimentellen Belegen.
New technologies in engineering, physics and biomedicine are demanding increasingly complex methods of digital signal processing. By presenting the latest research work the authors demonstrate how real-time recurrent neural networks (RNNs) can be implemented to expand the range of traditional signal processing techniques and to help combat the problem of prediction. Within this text neural networks are considered as massively interconnected nonlinear adaptive filters. * Analyses the relationships between RNNs and various nonlinear models and filters, and introduces spatio-temporal architectures together with the concepts of modularity and nesting * Examines stability and relaxation within RNNs * Presents on-line learning algorithms for nonlinear adaptive filters and introduces new paradigms which exploit the concepts of a priori and a posteriori errors, data-reusing adaptation, and normalisation * Studies convergence and stability of on-line learning algorithms based upon optimisation techniques such as contraction mapping and fixed point iteration * Describes strategies for the exploitation of inherent relationships between parameters in RNNs * Discusses practical issues such as predictability and nonlinearity detecting and includes several practical applications in areas such as air pollutant modelling and prediction, attractor discovery and chaos, ECG signal processing, and speech processing Recurrent Neural Networks for Prediction offers a new insight into the learning algorithms, architectures and stability of recurrent neural networks and, consequently, will have instant appeal. It provides an extensive background for researchers, academics and postgraduates enabling them to apply such networks in new applications.
Autorenporträt
Danilo Mandic from the Imperial College London, London, UK was named Fellow of the Institute of Electrical and Electronics Engineers in 2013 for contributions to multivariate and nonlinear learning systems. Jonathon A. Chambers is the author of Recurrent Neural Networks for Prediction: Learning Algorithms, Architectures and Stability, published by Wiley.