Normal view MARC view ISBD view

Circuit complexity and neural networks / Ian Parberry.

By: Parberry, Ian [author.].
Contributor(s): IEEE Xplore (Online Service) [distributor.] | MIT Press [publisher.].
Material type: materialTypeLabelBookSeries: Foundations of computing: Publisher: Cambridge, Massachusetts : MIT Press, c1994Distributor: [Piscataqay, New Jersey] : IEEE Xplore, [1994]Description: 1 PDF (xxix, 270 pages) : illustrations.Content type: text Media type: electronic Carrier type: online resourceISBN: 9780262281249.Subject(s): Logic circuits | Computational complexity | Neural networks (Computer science)Genre/Form: Electronic books.Additional physical formats: Print version: No titleOnline resources: Abstract with links to resource Also available in print.Summary: Neural networks usually work adequately on small problems but can run into trouble when they are scaled up to problems involving large amounts of input data. Circuit Complexity and Neural Networks addresses the important question of how well neural networks scale - that is, how fast the computation time and number of neurons grow as the problem size increases. It surveys recent research in circuit complexity (a robust branch of theoretical computer science) and applies this work to a theoretical understanding of the problem of scalability.Most research in neural networks focuses on learning, yet it is important to understand the physical limitations of the network before the resources needed to solve a certain problem can be calculated. One of the aims of this book is to compare the complexity of neural networks and the complexity of conventional computers, looking at the computational ability and resources (neurons and time) that are a necessary part of the foundations of neural network learning.Circuit Complexity and Neural Networks contains a significant amount of background material on conventional complexity theory that will enable neural network scientists to learn about how complexity theory applies to their discipline, and allow complexity theorists to see how their discipline applies to neural networks.
    average rating: 0.0 (0 votes)
No physical items for this record

Includes bibliographical references (p. [251]-257) and index.

Restricted to subscribers or individual electronic text purchasers.

Neural networks usually work adequately on small problems but can run into trouble when they are scaled up to problems involving large amounts of input data. Circuit Complexity and Neural Networks addresses the important question of how well neural networks scale - that is, how fast the computation time and number of neurons grow as the problem size increases. It surveys recent research in circuit complexity (a robust branch of theoretical computer science) and applies this work to a theoretical understanding of the problem of scalability.Most research in neural networks focuses on learning, yet it is important to understand the physical limitations of the network before the resources needed to solve a certain problem can be calculated. One of the aims of this book is to compare the complexity of neural networks and the complexity of conventional computers, looking at the computational ability and resources (neurons and time) that are a necessary part of the foundations of neural network learning.Circuit Complexity and Neural Networks contains a significant amount of background material on conventional complexity theory that will enable neural network scientists to learn about how complexity theory applies to their discipline, and allow complexity theorists to see how their discipline applies to neural networks.

Also available in print.

Mode of access: World Wide Web

Description based on PDF viewed 12/23/2015.

There are no comments for this item.

Log in to your account to post a comment.