probably approximately correct learning
Recently Published Documents


TOTAL DOCUMENTS

12
(FIVE YEARS 1)

H-INDEX

3
(FIVE YEARS 0)

2011 ◽  
pp. 805-805
Author(s):  
Thomas Zeugmann ◽  
Pascal Poupart ◽  
James Kennedy ◽  
Xin Jin ◽  
Jiawei Han ◽  
...  

1995 ◽  
Vol 7 (5) ◽  
pp. 1040-1053 ◽  
Author(s):  
Wee Sun Lee ◽  
Peter L. Bartlett ◽  
Robert C. Williamson

We examine the relationship between the VC dimension and the number of parameters of a threshold smoothly parameterized function class. We show that the VC dimension of such a function class is at least k if there exists a k-dimensional differentiable manifold in the parameter space such that each member of the manifold corresponds to a different decision boundary. Using this result, we are able to obtain lower bounds on the VC dimension proportional to the number of parameters for several thresholded function classes including two-layer neural networks with certain smooth activation functions and radial basis functions with a gaussian basis. These lower bounds hold even if the magnitudes of the parameters are restricted to be arbitrarily small. In Valiant's probably approximately correct learning framework, this implies that the number of examples necessary for learning these function classes is at least linear in the number of parameters.


Sign in / Sign up

Export Citation Format

Share Document