One of the most challenging and fascinating problems of the theory of neural nets is that of asymptotic behavior, of how a system behaves as time proceeds. This is of particular relevance to many practical applications. Here we focus on association, generalization, and representation. We turn to the last topic first. The introductory chapter, "Global Analysis of Recurrent Neural Net- works," by Andreas Herz presents an in-depth analysis of how to construct a Lyapunov function for various types of dynamics and neural coding. It includes a review of the recent work with John Hopfield on integrate-and- fire neurons with local interactions. The chapter, "Receptive Fields and Maps in the Visual Cortex: Models of Ocular Dominance and Orientation Columns" by Ken Miller, explains how the primary visual cortex may asymptotically gain its specific structure through a self-organization process based on Hebbian learning. His argu- ment since has been shown to be rather susceptible to generalization.
Publisher: Springer-Verlag New York Inc.
You may also be interested in...
Mixed media product
Simply reserve online and pay at the counter when you collect. Available in shop from just two hours, subject to availability.
Thank you for your reservation
Your order is now being processed and we have sent a confirmation email to you at
When will my order be ready to collect?
Following the initial email, you will be contacted by the shop to confirm that your item is available for collection.
Call us on or send us an email at
Unfortunately there has been a problem with your order
Please try again or alternatively you can contact your chosen shop on or send us an email at