Kernels for Vector-Valued Functions: a Review

Mauricio A. ÁlvarezLorenzo RosascoNeil D. Lawrence
, 2011.

Abstract

Kernel methods are among the most popular techniques in machine learning. From a frequentist/discriminative perspective they play a central role in regularization theory as they provide a natural choice for the hypotheses space and the regularization functional through the notion of reproducing kernel Hilbert spaces. From a Bayesian/generative perspective they are the key in the context of Gaussian processes, where the kernel function is also known as the covariance function. Traditionally, kernel methods have been used in supervised learning problem with scalar outputs and indeed there has been a considerable amount of work devoted to designing and learning kernels. More recently there has been an increasing interest in methods that deal with multiple outputs, motivated partly by frameworks like multitask learning. In this paper, we review different methods to design or learn valid kernel functions for multiple outputs, paying particular attention to the connection between probabilistic and functional methods.

Cite this Paper


BibTeX
@Misc{Alvarez:kernels11, title = {Kernels for Vector-Valued Functions: a Review}, author = {Álvarez, Mauricio A. and Rosasco, Lorenzo and Lawrence, Neil D.}, year = {2011}, pdf = {http://arxiv.org/pdf/1106.6251}, url = {http://inverseprobability.com/publications/alvarez-kernels11.html}, abstract = {Kernel methods are among the most popular techniques in machine learning. From a frequentist/discriminative perspective they play a central role in regularization theory as they provide a natural choice for the hypotheses space and the regularization functional through the notion of reproducing kernel Hilbert spaces. From a Bayesian/generative perspective they are the key in the context of Gaussian processes, where the kernel function is also known as the covariance function. Traditionally, kernel methods have been used in supervised learning problem with scalar outputs and indeed there has been a considerable amount of work devoted to designing and learning kernels. More recently there has been an increasing interest in methods that deal with multiple outputs, motivated partly by frameworks like multitask learning. In this paper, we review different methods to design or learn valid kernel functions for multiple outputs, paying particular attention to the connection between probabilistic and functional methods.} }
Endnote
%0 Generic %T Kernels for Vector-Valued Functions: a Review %A Mauricio A. Álvarez %A Lorenzo Rosasco %A Neil D. Lawrence %D 2011 %F Alvarez:kernels11 %U http://inverseprobability.com/publications/alvarez-kernels11.html %X Kernel methods are among the most popular techniques in machine learning. From a frequentist/discriminative perspective they play a central role in regularization theory as they provide a natural choice for the hypotheses space and the regularization functional through the notion of reproducing kernel Hilbert spaces. From a Bayesian/generative perspective they are the key in the context of Gaussian processes, where the kernel function is also known as the covariance function. Traditionally, kernel methods have been used in supervised learning problem with scalar outputs and indeed there has been a considerable amount of work devoted to designing and learning kernels. More recently there has been an increasing interest in methods that deal with multiple outputs, motivated partly by frameworks like multitask learning. In this paper, we review different methods to design or learn valid kernel functions for multiple outputs, paying particular attention to the connection between probabilistic and functional methods.
RIS
TY - GEN TI - Kernels for Vector-Valued Functions: a Review AU - Mauricio A. Álvarez AU - Lorenzo Rosasco AU - Neil D. Lawrence DA - 2011/06/30 ID - Alvarez:kernels11 L1 - http://arxiv.org/pdf/1106.6251 UR - http://inverseprobability.com/publications/alvarez-kernels11.html AB - Kernel methods are among the most popular techniques in machine learning. From a frequentist/discriminative perspective they play a central role in regularization theory as they provide a natural choice for the hypotheses space and the regularization functional through the notion of reproducing kernel Hilbert spaces. From a Bayesian/generative perspective they are the key in the context of Gaussian processes, where the kernel function is also known as the covariance function. Traditionally, kernel methods have been used in supervised learning problem with scalar outputs and indeed there has been a considerable amount of work devoted to designing and learning kernels. More recently there has been an increasing interest in methods that deal with multiple outputs, motivated partly by frameworks like multitask learning. In this paper, we review different methods to design or learn valid kernel functions for multiple outputs, paying particular attention to the connection between probabilistic and functional methods. ER -
APA
Álvarez, M.A., Rosasco, L. & Lawrence, N.D.. (2011). Kernels for Vector-Valued Functions: a Review. Available from http://inverseprobability.com/publications/alvarez-kernels11.html.

Related Material