HOME

TheInfoList



OR:

In computational learning theory, probably approximately correct (PAC) learning is a framework for mathematical analysis of
machine learning Machine learning (ML) is a field of inquiry devoted to understanding and building methods that 'learn', that is, methods that leverage data to improve performance on some set of tasks. It is seen as a part of artificial intelligence. Machine ...
. It was proposed in 1984 by
Leslie Valiant Leslie Gabriel Valiant (born 28 March 1949) is a British American computer scientist and computational theorist. He was born to a chemical engineer father and a translator mother. He is currently the T. Jefferson Coolidge Professor of Comput ...
.L. Valiant.
A theory of the learnable.
' Communications of the ACM, 27, 1984.
In this framework, the learner receives samples and must select a generalization function (called the ''hypothesis'') from a certain class of possible functions. The goal is that, with high probability (the "probably" part), the selected function will have low
generalization error For supervised learning applications in machine learning and statistical learning theory, generalization error (also known as the out-of-sample error or the risk) is a measure of how accurately an algorithm is able to predict outcome values for p ...
(the "approximately correct" part). The learner must be able to learn the concept given any arbitrary approximation ratio, probability of success, or distribution of the samples. The model was later extended to treat noise (misclassified samples). An important innovation of the PAC framework is the introduction of
computational complexity theory In theoretical computer science and mathematics, computational complexity theory focuses on classifying computational problems according to their resource usage, and relating these classes to each other. A computational problem is a task solved ...
concepts to machine learning. In particular, the learner is expected to find efficient functions (time and space requirements bounded to a
polynomial In mathematics, a polynomial is an expression consisting of indeterminates (also called variables) and coefficients, that involves only the operations of addition, subtraction, multiplication, and positive-integer powers of variables. An exampl ...
of the example size), and the learner itself must implement an efficient procedure (requiring an example count bounded to a polynomial of the concept size, modified by the approximation and
likelihood The likelihood function (often simply called the likelihood) represents the probability of random variable realizations conditional on particular values of the statistical parameters. Thus, when evaluated on a given sample, the likelihood functi ...
bounds).


Definitions and terminology

In order to give the definition for something that is PAC-learnable, we first have to introduce some terminology. For the following definitions, two examples will be used. The first is the problem of
character recognition Optical character recognition or optical character reader (OCR) is the electronic or mechanical conversion of images of typed, handwritten or printed text into machine-encoded text, whether from a scanned document, a photo of a document, a scen ...
given an array of n bits encoding a binary-valued image. The other example is the problem of finding an interval that will correctly classify points within the interval as positive and the points outside of the range as negative. Let X be a set called the ''instance space'' or the encoding of all the samples. In the character recognition problem, the instance space is X=\^n. In the interval problem the instance space, X, is the set of all bounded intervals in \mathbb, where \mathbb denotes the set of all
real numbers In mathematics, a real number is a number that can be used to measure a ''continuous'' one-dimensional quantity such as a distance, duration or temperature. Here, ''continuous'' means that values can have arbitrarily small variations. Every re ...
. A ''concept'' is a subset c \subset X. One concept is the set of all patterns of bits in X=\^n that encode a picture of the letter "P". An example concept from the second example is the set of open intervals, \, each of which contains only the positive points. A ''
concept class In computational learning theory in mathematics, a concept over a domain ''X'' is a total Boolean function over ''X''. A concept class is a class of concepts. Concept classes are a subject of computational learning theory. Concept class terminolog ...
'' C is a collection of concepts over X. This could be the set of all subsets of the array of bits that are skeletonized 4-connected (width of the font is 1). Let EX(c,D) be a procedure that draws an example, x, using a probability distribution D and gives the correct label c(x), that is 1 if x \in c and 0 otherwise. Now, given 0<\epsilon,\delta<1 , assume there is an algorithm A and a polynomial p in 1/\epsilon, 1/\delta (and other relevant parameters of the class C) such that, given a sample of size p drawn according to EX(c,D), then, with probability of at least 1-\delta, A outputs a hypothesis h \in C that has an average error less than or equal to \epsilon on X with the same distribution D. Further if the above statement for algorithm A is true for every concept c \in C and for every distribution D over X, and for all 0<\epsilon, \delta<1 then C is (efficiently) PAC learnable (or ''distribution-free PAC learnable''). We can also say that A is a PAC learning algorithm for C.


Equivalence

Under some regularity conditions these conditions are equivalent: # The concept class ''C'' is PAC learnable. # The
VC dimension VC may refer to: Military decorations * Victoria Cross, a military decoration awarded by the United Kingdom and also by certain Commonwealth nations ** Victoria Cross for Australia ** Victoria Cross (Canada) ** Victoria Cross for New Zealand * Vic ...
of ''C'' is finite. # ''C'' is a uniform Glivenko–Cantelli class. # ''C'' is compressible in the sense of Littlestone and Warmuth


See also

*
Occam learning In computational learning theory, Occam learning is a model of algorithmic learning where the objective of the learner is to output a succinct representation of received training data. This is closely related to probably approximately correct (P ...
* Data mining * Error tolerance (PAC learning) *
Sample complexity The sample complexity of a machine learning algorithm represents the number of training-samples that it needs in order to successfully learn a target function. More precisely, the sample complexity is the number of training-samples that we need to ...


References


Further reading

* M. Kearns, U. Vazirani.
An Introduction to Computational Learning Theory
'' MIT Press, 1994. A textbook. * M. Mohri, A. Rostamizadeh, and A. Talwalkar. ''Foundations of Machine Learning''. MIT Press, 2018. Chapter 2 contains a detailed treatment of PAC-learnability
Readable through open access from the publisher.
* D. Haussler
Overview of the Probably Approximately Correct (PAC) Learning Framework
An introduction to the topic. * L. Valiant
''Probably Approximately Correct.''
Basic Books, 2013. In which Valiant argues that PAC learning describes how organisms evolve and learn. * * {{cite arXiv, eprint=1503.06960, last1=Moran, first1=Shay, last2=Yehudayoff, first2=Amir, title=Sample compression schemes for VC classes, year=2015, class=cs.LG Computational learning theory