HOME

TheInfoList



OR:

In machine learning, kernel machines are a class of algorithms for
pattern analysis Pattern recognition is the automated recognition of patterns and regularities in data. It has applications in statistical data analysis, signal processing, image analysis, information retrieval, bioinformatics, data compression, computer graphics ...
, whose best known member is the
support-vector machine In machine learning, support vector machines (SVMs, also support vector networks) are supervised learning models with associated learning algorithms that analyze data for classification and regression analysis. Developed at AT&T Bell Laboratorie ...
(SVM). The general task of
pattern analysis Pattern recognition is the automated recognition of patterns and regularities in data. It has applications in statistical data analysis, signal processing, image analysis, information retrieval, bioinformatics, data compression, computer graphics ...
is to find and study general types of relations (for example clusters, rankings,
principal components Principal component analysis (PCA) is a popular technique for analyzing large datasets containing a high number of dimensions/features per observation, increasing the interpretability of data while preserving the maximum amount of information, and ...
,
correlation In statistics, correlation or dependence is any statistical relationship, whether causal or not, between two random variables or bivariate data. Although in the broadest sense, "correlation" may indicate any type of association, in statistics ...
s, classifications) in datasets. For many algorithms that solve these tasks, the data in raw representation have to be explicitly transformed into feature vector representations via a user-specified ''feature map'': in contrast, kernel methods require only a user-specified ''kernel'', i.e., a similarity function over all pairs of data points computed using
Inner products In mathematics, an inner product space (or, rarely, a Hausdorff pre-Hilbert space) is a real vector space or a complex vector space with an operation called an inner product. The inner product of two vectors in the space is a scalar, often d ...
. The feature map in kernel machines is infinite dimensional but only requires a finite dimensional matrix from user-input according to the Representer theorem. Kernel machines are slow to compute for datasets larger than a couple of thousand examples without parallel processing. Kernel methods owe their name to the use of kernel functions, which enable them to operate in a high-dimensional, ''implicit'' feature space without ever computing the coordinates of the data in that space, but rather by simply computing the inner products between the images of all pairs of data in the feature space. This operation is often computationally cheaper than the explicit computation of the coordinates. This approach is called the "kernel trick". Kernel functions have been introduced for sequence data, graphs, text, images, as well as vectors. Algorithms capable of operating with kernels include the kernel perceptron, support-vector machines (SVM),
Gaussian process In probability theory and statistics, a Gaussian process is a stochastic process (a collection of random variables indexed by time or space), such that every finite collection of those random variables has a multivariate normal distribution, i.e. e ...
es, principal components analysis (PCA), canonical correlation analysis, ridge regression, spectral clustering, linear adaptive filters and many others. Most kernel algorithms are based on convex optimization or eigenproblems and are statistically well-founded. Typically, their statistical properties are analyzed using statistical learning theory (for example, using Rademacher complexity).


Motivation and informal explanation

Kernel methods can be thought of as instance-based learners: rather than learning some fixed set of parameters corresponding to the features of their inputs, they instead "remember" the i-th training example (\mathbf_i, y_i) and learn for it a corresponding weight w_i. Prediction for unlabeled inputs, i.e., those not in the training set, is treated by the application of a similarity function k, called a kernel, between the unlabeled input \mathbf and each of the training inputs \mathbf_i. For instance, a kernelized
binary classifier Binary classification is the task of classifying the elements of a set into two groups (each called ''class'') on the basis of a classification rule. Typical binary classification problems include: * Medical testing to determine if a patient has c ...
typically computes a weighted sum of similarities :\hat = \sgn \sum_^n w_i y_i k(\mathbf_i, \mathbf), where * \hat \in \ is the kernelized binary classifier's predicted label for the unlabeled input \mathbf whose hidden true label y is of interest; * k \colon \mathcal \times \mathcal \to \mathbb is the kernel function that measures similarity between any pair of inputs \mathbf, \mathbf \in \mathcal; * the sum ranges over the labeled examples \_^n in the classifier's training set, with y_i \in \; * the w_i \in \mathbb are the weights for the training examples, as determined by the learning algorithm; * the
sign function In mathematics, the sign function or signum function (from '' signum'', Latin for "sign") is an odd mathematical function that extracts the sign of a real number. In mathematical expressions the sign function is often represented as . To avoi ...
\sgn determines whether the predicted classification \hat comes out positive or negative. Kernel classifiers were described as early as the 1960s, with the invention of the kernel perceptron. They rose to great prominence with the popularity of the
support-vector machine In machine learning, support vector machines (SVMs, also support vector networks) are supervised learning models with associated learning algorithms that analyze data for classification and regression analysis. Developed at AT&T Bell Laboratorie ...
(SVM) in the 1990s, when the SVM was found to be competitive with
neural networks A neural network is a network or circuit of biological neurons, or, in a modern sense, an artificial neural network, composed of artificial neurons or nodes. Thus, a neural network is either a biological neural network, made up of biological ...
on tasks such as handwriting recognition.


Mathematics: the kernel trick

The kernel trick avoids the explicit mapping that is needed to get linear
learning algorithms Machine learning (ML) is a field of inquiry devoted to understanding and building methods that 'learn', that is, methods that leverage data to improve performance on some set of tasks. It is seen as a part of artificial intelligence. Machine ...
to learn a nonlinear function or decision boundary. For all \mathbf and \mathbf in the input space \mathcal, certain functions k(\mathbf, \mathbf) can be expressed as an inner product in another space \mathcal. The function k \colon \mathcal \times \mathcal \to \mathbb is often referred to as a ''kernel'' or a '' kernel function''. The word "kernel" is used in mathematics to denote a weighting function for a weighted sum or integral. Certain problems in machine learning have more structure than an arbitrary weighting function k. The computation is made much simpler if the kernel can be written in the form of a "feature map" \varphi\colon \mathcal \to \mathcal which satisfies :k(\mathbf, \mathbf) = \langle \varphi(\mathbf), \varphi(\mathbf) \rangle_\mathcal. The key restriction is that \langle \cdot, \cdot \rangle_\mathcal must be a proper inner product. On the other hand, an explicit representation for \varphi is not necessary, as long as \mathcal is an inner product space. The alternative follows from Mercer's theorem: an implicitly defined function \varphi exists whenever the space \mathcal can be equipped with a suitable measure ensuring the function k satisfies Mercer's condition. Mercer's theorem is similar to a generalization of the result from linear algebra that associates an inner product to any positive-definite matrix. In fact, Mercer's condition can be reduced to this simpler case. If we choose as our measure the counting measure \mu(T) = , T, for all T \subset X , which counts the number of points inside the set T, then the integral in Mercer's theorem reduces to a summation : \sum_^n\sum_^n k(\mathbf_i, \mathbf_j) c_i c_j \geq 0. If this summation holds for all finite sequences of points (\mathbf_1, \dotsc, \mathbf_n) in \mathcal and all choices of n real-valued coefficients (c_1, \dots, c_n) (cf.
positive definite kernel In operator theory, a branch of mathematics, a positive-definite kernel is a generalization of a positive-definite function or a positive-definite matrix. It was first introduced by James Mercer (mathematician), James Mercer in the early 20th centur ...
), then the function k satisfies Mercer's condition. Some algorithms that depend on arbitrary relationships in the native space \mathcal would, in fact, have a linear interpretation in a different setting: the range space of \varphi. The linear interpretation gives us insight about the algorithm. Furthermore, there is often no need to compute \varphi directly during computation, as is the case with
support-vector machine In machine learning, support vector machines (SVMs, also support vector networks) are supervised learning models with associated learning algorithms that analyze data for classification and regression analysis. Developed at AT&T Bell Laboratorie ...
s. Some cite this running time shortcut as the primary benefit. Researchers also use it to justify the meanings and properties of existing algorithms. Theoretically, a Gram matrix \mathbf \in \mathbb^ with respect to \ (sometimes also called a "kernel matrix"), where K_ = k(\mathbf_i, \mathbf_j), must be positive semi-definite (PSD). Empirically, for machine learning heuristics, choices of a function k that do not satisfy Mercer's condition may still perform reasonably if k at least approximates the intuitive idea of similarity. Regardless of whether k is a Mercer kernel, k may still be referred to as a "kernel". If the kernel function k is also a covariance function as used in
Gaussian processes In probability theory and statistics, a Gaussian process is a stochastic process (a collection of random variables indexed by time or space), such that every finite collection of those random variables has a multivariate normal distribution, i.e. ...
, then the Gram matrix \mathbf can also be called a
covariance matrix In probability theory and statistics, a covariance matrix (also known as auto-covariance matrix, dispersion matrix, variance matrix, or variance–covariance matrix) is a square matrix giving the covariance between each pair of elements of ...
.


Applications

Application areas of kernel methods are diverse and include geostatistics, kriging, inverse distance weighting,
3D reconstruction In computer vision and computer graphics, 3D reconstruction is the process of capturing the shape and appearance of real objects. This process can be accomplished either by active or passive methods. If the model is allowed to change its shape i ...
,
bioinformatics Bioinformatics () is an interdisciplinary field that develops methods and software tools for understanding biological data, in particular when the data sets are large and complex. As an interdisciplinary field of science, bioinformatics combi ...
, chemoinformatics, information extraction and handwriting recognition.


Popular kernels

* Fisher kernel * Graph kernels * Kernel smoother * Polynomial kernel * Radial basis function kernel (RBF) *
String kernel In machine learning and data mining, a string kernel is a kernel function that operates on strings, i.e. finite sequences of symbols that need not be of the same length. String kernels can be intuitively understood as functions measuring the simila ...
s *
Neural tangent kernel In the study of artificial neural networks (ANNs), the neural tangent kernel (NTK) is a kernel that describes the evolution of deep artificial neural networks during their training by gradient descent. It allows ANNs to be studied using theoretica ...
* Neural network Gaussian process (NNGP) kernel


See also

* Kernel methods for vector output * Kernel density estimation * Representer theorem *
Similarity learning Similarity learning is an area of supervised machine learning in artificial intelligence. It is closely related to regression and classification, but the goal is to learn a similarity function that measures how similar or related two objects are ...
*
Cover's theorem Cover's theorem is a statement in computational learning theory and is one of the primary theoretical motivations for the use of non-linear kernel methods in machine learning applications. It is so termed after the information theorist Thomas M. Cov ...


References


Further reading

* * *


External links


Kernel-Machines Org
��community website
onlineprediction.net Kernel Methods Article
{{DEFAULTSORT:Kernel Methods Kernel methods for machine learning Geostatistics Classification algorithms