Definition
Probably Approximately Correct Learning
The Probably Approximately Correct (PAC) learning framework is a model in computational learning theory that analyzes whether and how efficiently a learning algorithm can produce a model with good generalisation performance. It provides a formal guarantee that the learned model will be mostly accurate, most of the time.
The name breaks down its core guarantee for a learned hypothesis :
- Approximately Correct: The model’s true error (risk) is low, bounded by a small error parameter .
- Probably: This guarantee of being “approximately correct” holds with a high probability, at least , where is a small confidence parameter.
Formalism
A hypothesis space is considered PAC-learnable if, for any choice of and , there is a learning algorithm that, given a sufficient number of i.i.d. training samples, will produce a hypothesis satisfying:
The central aim of the PAC framework is to determine the sample complexity—the minimum number of training samples required to meet these guarantees for a given hypothesis space. This links the amount of data needed to the desired level of accuracy and confidence.