Asset Details
MbrlCatalogueTitleDetail
Do you wish to reserve the book?
Noise-Adaptive Confidence Sets for Linear Bandits and Application to Bayesian Optimization
by
Kim, Jungtaek
, Kwang-Sung, Jun
in
Adaptive algorithms
/ Bayesian analysis
/ Confidence
/ Decision theory
/ Empirical analysis
/ Noise levels
/ Optimization
/ Parameters
2024
Hey, we have placed the reservation for you!
By the way, why not check out events that you can attend while you pick your title.
You are currently in the queue to collect this book. You will be notified once it is your turn to collect the book.
Oops! Something went wrong.
Looks like we were not able to place the reservation. Kindly try again later.
Are you sure you want to remove the book from the shelf?
Oops! Something went wrong.
While trying to remove the title from your shelf something went wrong :( Kindly try again later!
Do you wish to request the book?
Noise-Adaptive Confidence Sets for Linear Bandits and Application to Bayesian Optimization
by
Kim, Jungtaek
, Kwang-Sung, Jun
in
Adaptive algorithms
/ Bayesian analysis
/ Confidence
/ Decision theory
/ Empirical analysis
/ Noise levels
/ Optimization
/ Parameters
2024
Please be aware that the book you have requested cannot be checked out. If you would like to checkout this book, you can reserve another copy
We have requested the book for you!
Your request is successful and it will be processed during the Library working hours. Please check the status of your request in My Requests.
Oops! Something went wrong.
Looks like we were not able to place your request. Kindly try again later.
Noise-Adaptive Confidence Sets for Linear Bandits and Application to Bayesian Optimization
Paper
Noise-Adaptive Confidence Sets for Linear Bandits and Application to Bayesian Optimization
2024
Request Book From Autostore
and Choose the Collection Method
Overview
Adapting to a priori unknown noise level is a very important but challenging problem in sequential decision-making as efficient exploration typically requires knowledge of the noise level, which is often loosely specified. We report significant progress in addressing this issue for linear bandits in two respects. First, we propose a novel confidence set that is `semi-adaptive' to the unknown sub-Gaussian parameter \\(\\sigma_*^2\\) in the sense that the (normalized) confidence width scales with \\(\\sqrt{d\\sigma_*^2 + \\sigma_0^2}\\) where \\(d\\) is the dimension and \\(\\sigma_0^2\\) is the specified sub-Gaussian parameter (known) that can be much larger than \\(\\sigma_*^2\\). This is a significant improvement over \\(\\sqrt{d\\sigma_0^2}\\) of the standard confidence set of Abbasi-Yadkori et al. (2011), especially when \\(d\\) is large or \\(\\sigma_*^2=0\\). We show that this leads to an improved regret bound in linear bandits. Second, for bounded rewards, we propose a novel variance-adaptive confidence set that has much improved numerical performance upon prior art. We then apply this confidence set to develop, as we claim, the first practical variance-adaptive linear bandit algorithm via an optimistic approach, which is enabled by our novel regret analysis technique. Both of our confidence sets rely critically on `regret equality' from online learning. Our empirical evaluation in diverse Bayesian optimization tasks shows that our proposed algorithms demonstrate better or comparable performance compared to existing methods.
Publisher
Cornell University Library, arXiv.org
Subject
This website uses cookies to ensure you get the best experience on our website.