• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations
Advanced Search Include Citations

DMCA

Gaussian Process Optimization in the Bandit Setting: No Regret and Experimental Design

Cached

  • Download as a PDF

Download Links

  • [www.cs.caltech.edu]
  • [www.icml2010.org]
  • [www.cs.cmu.edu]
  • [www-2.cs.cmu.edu]
  • [icml2010.haifa.il.ibm.com]
  • [users.cms.caltech.edu]
  • [users.cms.caltech.edu]
  • [homes.cs.washington.edu]
  • [stat.wharton.upenn.edu]
  • [users.cms.caltech.edu]
  • [arxiv.org]

  • Save to List
  • Add to Collection
  • Correct Errors
  • Monitor Changes
by Niranjan Srinivas , Andreas Krause , Matthias Seeger
Citations:125 - 13 self
  • Summary
  • Citations
  • Active Bibliography
  • Co-citation
  • Clustered Documents
  • Version History

BibTeX

@MISC{Srinivas_gaussianprocess,
    author = {Niranjan Srinivas and Andreas Krause and Matthias Seeger},
    title = {Gaussian Process Optimization in the Bandit Setting: No Regret and Experimental Design},
    year = {}
}

Share

Facebook Twitter Reddit Bibsonomy

OpenURL

 

Abstract

Many applications require optimizing an unknown, noisy function that is expensive to evaluate. We formalize this task as a multiarmed bandit problem, where the payoff function is either sampled from a Gaussian process (GP) or has low RKHS norm. We resolve the important open problem of deriving regret bounds for this setting, which imply novel convergence rates for GP optimization. We analyze GP-UCB, an intuitive upper-confidence based algorithm, and bound its cumulative regret in terms of maximal information gain, establishing a novel connection between GP optimization and experimental design. Moreover, by bounding the latter in terms of operator spectra, we obtain explicit sublinear regret bounds for many commonly used covariance functions. In some important cases, our bounds have surprisingly weak dependence on the dimensionality. In our experiments on real sensor data, GP-UCB compares favorably with other heuristical GP optimization approaches. 1.

Keyphrases

experimental design    bandit setting    gaussian process optimization    gp optimization    cumulative regret    gaussian process    many application    payoff function    covariance function    important case    novel convergence rate    important open problem    explicit sublinear regret bound    novel connection    regret bound    real sensor data    noisy function    maximal information gain    operator spectrum    intuitive upper-confidence    weak dependence    heuristical gp optimization approach    low rkhs norm    multiarmed bandit problem   

Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University