National Science Library of Georgia

Image from Google Jackets

Neural network learning : theoretical foundations / Martin Anthony and Peter L. Bartlett.

By: Contributor(s): Material type: TextTextPublisher: Cambridge : Cambridge University Press, 1999Description: 1 online resource (xiv, 389 pages) : digital, PDF file(s)Content type:
  • text
Media type:
  • computer
Carrier type:
  • online resource
ISBN:
  • 9780511624216 (ebook)
Subject(s): Additional physical formats: Print version: : No titleDDC classification:
  • 006.3/2 21
LOC classification:
  • QA76.87 .A58 1999
Online resources: Summary: This book describes theoretical advances in the study of artificial neural networks. It explores probabilistic models of supervised learning problems, and addresses the key statistical and computational questions. Research on pattern classification with binary-output networks is surveyed, including a discussion of the relevance of the Vapnik-Chervonenkis dimension, and calculating estimates of the dimension for several neural network models. A model of classification by real-output networks is developed, and the usefulness of classification with a 'large margin' is demonstrated. The authors explain the role of scale-sensitive versions of the Vapnik-Chervonenkis dimension in large margin classification, and in real prediction. They also discuss the computational complexity of neural network learning, describing a variety of hardness results, and outlining two efficient constructive learning algorithms. The book is self-contained and is intended to be accessible to researchers and graduate students in computer science, engineering, and mathematics.
Tags from this library: No tags from this library for this title. Log in to add tags.
No physical items for this record

Title from publisher's bibliographic system (viewed on 05 Oct 2015).

This book describes theoretical advances in the study of artificial neural networks. It explores probabilistic models of supervised learning problems, and addresses the key statistical and computational questions. Research on pattern classification with binary-output networks is surveyed, including a discussion of the relevance of the Vapnik-Chervonenkis dimension, and calculating estimates of the dimension for several neural network models. A model of classification by real-output networks is developed, and the usefulness of classification with a 'large margin' is demonstrated. The authors explain the role of scale-sensitive versions of the Vapnik-Chervonenkis dimension in large margin classification, and in real prediction. They also discuss the computational complexity of neural network learning, describing a variety of hardness results, and outlining two efficient constructive learning algorithms. The book is self-contained and is intended to be accessible to researchers and graduate students in computer science, engineering, and mathematics.

There are no comments on this title.

to post a comment.
Copyright © 2023 Sciencelib.ge All rights reserved.