machine-learning

Definition

Probably Approximately Correct Learning

The Probably Approximately Correct (PAC) learning framework is a model in computational learning theory that analyzes whether and how efficiently a learning algorithm can produce a model with good generalisation performance. It provides a formal guarantee that the learned model will be mostly accurate, most of the time.

The name breaks down its core guarantee for a learned hypothesis :

  • Approximately Correct: The model’s true error (risk) is low, bounded by a small error parameter .
  • Probably: This guarantee of being “approximately correct” holds with a high probability, at least , where is a small confidence parameter.

Formalism

A hypothesis space is considered PAC-learnable if, for any choice of and , there is a learning algorithm that, given a sufficient number of i.i.d. training samples, will produce a hypothesis satisfying:

The central aim of the PAC framework is to determine the sample complexity—the minimum number of training samples required to meet these guarantees for a given hypothesis space. This links the amount of data needed to the desired level of accuracy and confidence.