machine-learning learning-theory
Definition
Realisable PAC-Learnable Hypothesis Class
A hypothesis class is realisable PAC-learnable if there exists a learning algorithm and a sample complexity function such that for any parameters and any distribution satisfying the realisability assumption, the following holds:
provided that the sample size . The produced hypothesis is approximately correct (error ) with a probability of at least (probably).
Empirical Risk Minimisation
A hypothesis class is realisable PAC-learnable by any ERM algorithm provided that the functional capacity of the class is bounded. For finite classes, the sample complexity scales with the logarithm of the cardinality . For infinite classes, learnability is guaranteed if and only if the VC dimension is finite, with the complexity bounded by:
This result establishes that the complexity of the functional space is the fundamental constraint on the amount of data required to ensure generalisation.
Finite-class realisable PAC guarantee ( ERM)
Let be finite, , and assume the realisability assumption, i.e. there exists with . Let be an ERM algorithm and let for .
Here denotes the true risk of under distribution (equivalently when is implicit).Since , we have for every sample , hence ERM can return a consistent hypothesis and therefore .
Define the bad set
Fix any . Under binary classification with 0-1 loss,
where the second equality uses that the expectation of an indicator equals the probability of its event. Hence
Since implies , we obtain
By i.i.d. sampling,
Thus a fixed bad hypothesis is consistent with probability at most .
Now consider the failure event . Because is consistent, failure implies that at least one bad hypothesis is consistent with . Therefore, by the union bound,
To make this at most , it is sufficient that
Hence, for all satisfying this bound,
so finite is realisable PAC-learnable by ERM.