skip to main content
10.1145/1553374.1553492acmotherconferencesArticle/Chapter ViewAbstractPublication PagesicmlConference Proceedingsconference-collections
research-article

Function factorization using warped Gaussian processes

Published: 14 June 2009 Publication History

Abstract

We introduce a new approach to non-linear regression called function factorization, that is suitable for problems where an output variable can reasonably be modeled by a number of multiplicative interaction terms between non-linear functions of the inputs. The idea is to approximate a complicated function on a high-dimensional space by the sum of products of simpler functions on lower-dimensional subspaces. Function factorization can be seen as a generalization of matrix and tensor factorization methods, in which the data are approximated by the sum of outer products of vectors. We present a non-parametric Bayesian approach to function factorization where the priors over the factorizing functions are warped Gaussian processes, and we do inference using Hamiltonian Markov chain Monte Carlo. We demonstrate the superior predictive performance of the method on a food science data set compared to Gaussian process regression and tensor factorization using PARAFAC and GEMANOVA models.

References

[1]
Adams, R., & Stegle, O. (2008). Gaussian process product models for nonparametric nonstationarity. Machine Learning, International Conference on (ICML) (pp. 1--8).
[2]
Andersson, C., & Bro, R. (2000). The n-way toolbox for matlab. Chemometrics & Intelligent Laboratory Systems, 52, 1--4.
[3]
Bro, R., & Jakobsen, M. (2002). Exploring complex interactions in designed data using gemanova. color changes in fresh beef during storage. Chemometrics, Journal of, 16, 294--304.
[4]
Duane, S., Kennedy, A. D., Pendleton, B. J., & Roweth, D. (1987). Hybrid monte carlo. Physics Letters B, 195, 216--222.
[5]
Rasmussen, C. E., & Williams, C. K. I. (2006). Gaussian processes for machine learning. MIT Press.
[6]
Schmidt, M., & Laurberg, H. (2008). Nonnegative matrix factorization with gaussian process priors. Computational Intelligence and Neuroscience, 2008. 10.1155/2008/361705.
[7]
Smith, S. P. (1995). Differentiation of the cholesky algorithm. Computational and Graphical Statistics, Journal of, 4, 134--147.
[8]
Snelson, E., Rasmussen, C. E., & Ghahramani, Z. (2004). Warped gaussian processes. Neural Information Processing Systems, Advances in (NIPS) (pp. 337--344).
[9]
Tomasi, G., & Bro, R. (2002). Parafac and missing values. Chemometrics and Intelligent Laboratory Systems, 75, 163--180. 10.1016/j.chemolab.2004.07.003.

Cited By

View all

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Other conferences
ICML '09: Proceedings of the 26th Annual International Conference on Machine Learning
June 2009
1331 pages
ISBN:9781605585161
DOI:10.1145/1553374

Sponsors

  • NSF
  • Microsoft Research: Microsoft Research
  • MITACS

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 14 June 2009

Permissions

Request permissions for this article.

Check for updates

Qualifiers

  • Research-article

Conference

ICML '09
Sponsor:
  • Microsoft Research

Acceptance Rates

Overall Acceptance Rate 140 of 548 submissions, 26%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)14
  • Downloads (Last 6 weeks)2
Reflects downloads up to 03 Feb 2025

Other Metrics

Citations

Cited By

View all

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media