Efficient Multioutput Gaussian Processes through Variational Inducing Kernels.

Journal of Machine Learning Research - Proceedings Track 01/2010; 9:25-32.
Source: DBLP

ABSTRACT Interest in multioutput kernel methods is increas- ing, whether under the guise of multitask learn- ing, multisensor networks or structured output data. From the Gaussian process perspective a multioutput Mercer kernel is a covariance func- tion over correlated output functions. One way of constructing such kernels is based on convolution processes (CP). A key problem for this approach is efficient inference. ´ Alvarez and Lawrence re- cently presented a sparse approximation for CPs that enabled efficient inference. In this paper, we extend this work in two directions: we in- troduce the concept of variational inducing func- tions to handle potential non-smooth functions involved in the kernel CP construction and we consider an alternative approach to approximate inference based on variational methods, extend- ing the work by Titsias (2009) to the multiple output case. We demonstrate our approaches on prediction of school marks, compiler perfor- mance and financial time series.

  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: This paper proposes an online tree-based Bayesian approach for reinforcement learning. For inference, we employ a generalised context tree model. This defines a distribution on multivariate Gaussian piecewise-linear models, which can be updated in closed form. The tree structure itself is constructed using the cover tree method, which remains efficient in high dimensional spaces. We combine the model with Thompson sampling and approximate dynamic programming to obtain effective exploration policies in unknown environments. The flexibility and computational simplicity of the model render it suitable for many reinforcement learning problems in continuous state spaces. We demonstrate this in an experimental comparison with least squares policy iteration.
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: Multi-task learning models using Gaussian processes (GP) have been recently developed and successfully applied in various applications. The main difficulty with this approach is the computational cost of inference using the union of examples from all tasks. The paper investigates this problem for the grouped mixed-effect GP model where each individual response is given by a fixed-effect, taken from one of a set of unknown groups, plus a random individual effect function that captures variations among individuals. Such models have been widely used in previous work but no sparse solutions have been developed. The paper presents the first sparse solution for such problems, showing how the sparse approximation can be obtained by maximizing a variational lower bound on the marginal likelihood, generalizing ideas from single-task Gaussian processes to handle the mixed-effect model as well as grouping. Experiments using artificial and real data validate the approach showing that it can recover the performance of inference with the full sample, that it outperforms baseline methods, and that it outperforms state of the art sparse solutions for other multi-task GP formulations.
    Proceedings of the 2012 European conference on Machine Learning and Knowledge Discovery in Databases - Volume Part I; 09/2012
  • Source
    [Show abstract] [Hide abstract]
    ABSTRACT: Purely data driven approaches for machine learning present difficulties when data is scarce relative to the complexity of the model or when the model is forced to extrapolate. On the other hand, purely mechanistic approaches need to identify and specify all the interactions in the problem at hand (which may not be feasible) and still leave the issue of how to parameterize the system. In this paper, we present a hybrid approach using Gaussian processes and differential equations to combine data driven modelling with a physical model of the system. We show how different, physically-inspired, kernel functions can be developed through sensible, simple, mechanistic assumptions about the underlying system. The versatility of our approach is illustrated with three case studies from motion capture, computational biology and geostatistics.
    IEEE Transactions on Pattern Analysis and Machine Intelligence 05/2013; · 4.80 Impact Factor

Full-text (2 Sources)

Available from
May 28, 2014