|    Login    |    Register

An Introduction to Computational Learning Theory

(Hardback)


Publishing Details

Full Title:

An Introduction to Computational Learning Theory

Contributors:

By (Author) Michael J. Kearns
By (author) Umesh Vazirani

ISBN:

9780262111935

Publisher:

MIT Press Ltd

Imprint:

MIT Press

Publication Date:

15th August 1994

Country:

United States

Classifications

Readership:

Professional and Scholarly

Fiction/Non-fiction:

Non Fiction

Main Subject:
Dewey:

006.3

Physical Properties

Physical Format:

Hardback

Number of Pages:

222

Dimensions:

Width 178mm, Height 229mm, Spine 17mm

Weight:

590g

Description

Emphasizing issues of computational efficiency, this text introduces a number of central topics in computational learning theory for researchers and students in artificial intelligence, neural networks, theoretical computer science, and statistics. Computational learning theory is a new and rapidly expanding area of research that examines formal models of induction with the goals of discovering the common methods underlying efficient learning algorithms and identifying the computational impediments to learning. Each topic in the book has been chosen to elucidate a general principle, which is explored in a precise formal setting. Intuition has been emphasized in the presentation to make the material accessible to the nontheoretician while still providing precise arguments for the specialist. This balance is the result of new proofs of established theorems, and new presentations of the standard proofs. The topics covered include the motivation, definitions and fundamental results, both positive and negative, for the widely studied L.G. Valiant model of Probably Approximately Correct Learning; Occam's Razor, which formalizes a relationship between learning and data compression; the Vapnik-Chervonenkis dimension; the equivalence of weak and strong learning; efficient learning in the presence of noise by the method of statistical queries; relationships between learning and cryptography, and the resulting computational limitations on efficient learning; reducibility between learning problems; and algorithms for learning finite automata from active experimentation.

Author Bio

Michael J. Kearns is Professor of Computer and Information Science at the University of Pennsylvania. Umesh Vazirani is Roger A. Strauch Professor in the Electrical Engineering and Computer Sciences Department at the University of California, Berkeley.

See all

Other titles from MIT Press Ltd