This document discusses Gaussian process bandit optimization, which is a framework for adaptively sampling an unknown function to maximize information gain. It proposes using an Upper Confidence Bound (UCB) approach, where samples are selected to maximize an upper bound on the function value while balancing exploration and exploitation. The key results establish that the regret of this approach depends on the rate at which information can be gained about the function, which captures its "learnability." Experimental results on temperature and traffic data demonstrate the UCB approach performs comparably to existing heuristics.