Pac VC PDF
Pac VC PDF
Pac VC PDF
General: http://www.learning-with-kernels.org/
Example of more complex bounds:
http://www.research.ibm.com/people/t/tzhang/papers/jmlr02_cover.ps.gz
PAC-learning, VC
Dimension and
Margin-based Bounds
Machine Learning – 10701/15781
Carlos Guestrin
Carnegie Mellon University
Even if h makes zero errors in training data, may make errors in test
Using a PAC bound
Typically, 2 use cases:
1: Pick ε and δ, give you m
2: Pick m and δ, give you ε
Limitations of Haussler ‘88 bound
Consistent classifier
Write Lk = log2 Hk
L0 = 1
Lk+1 = log2 n + 2Lk
So Lk = (2k-1)(1+log2 n) +1
PAC bound for decision trees of
depth k
Bad!!!
Number of points is exponential in depth!
But, for m data points, decision tree can’t get too big…
Neural networks
VC(H) = #parameters
Local minima means NNs will probably not find best
parameters
1-Nearest neighbor?
PAC bound for SVMs
SVMs use a linear classifier
For d features, VC(H) = d+1:
VC dimension and SVMs: Problems!!!
Doesn’t take margin into account
n – input features
p – degree of polynomial
VC(H) = R2 w.w
R2 = maxj Φ(xj).Φ(xj) – magnitude of data, doesn’t depend on choice of w
SVMs minimize w.w
ε
d=2000
d=200
d=20
d=2
m (in 105)