Asset Details
MbrlCatalogueTitleDetail
Do you wish to reserve the book?
Variance-Reduced Accelerated First-order Methods: Central Limit Theorems and Confidence Statements
by
Lei, Jinlong
, Shanbhag, Uday V
in
Confidence
/ Convergence
/ Convexity
/ Covariance matrix
/ Hessian matrices
/ Normal distribution
/ Optimization
/ Parameter estimation
/ Polynomials
/ Theorems
2024
Hey, we have placed the reservation for you!
By the way, why not check out events that you can attend while you pick your title.
You are currently in the queue to collect this book. You will be notified once it is your turn to collect the book.
Oops! Something went wrong.
Looks like we were not able to place the reservation. Kindly try again later.
Are you sure you want to remove the book from the shelf?
Variance-Reduced Accelerated First-order Methods: Central Limit Theorems and Confidence Statements
by
Lei, Jinlong
, Shanbhag, Uday V
in
Confidence
/ Convergence
/ Convexity
/ Covariance matrix
/ Hessian matrices
/ Normal distribution
/ Optimization
/ Parameter estimation
/ Polynomials
/ Theorems
2024
Oops! Something went wrong.
While trying to remove the title from your shelf something went wrong :( Kindly try again later!
Do you wish to request the book?
Variance-Reduced Accelerated First-order Methods: Central Limit Theorems and Confidence Statements
by
Lei, Jinlong
, Shanbhag, Uday V
in
Confidence
/ Convergence
/ Convexity
/ Covariance matrix
/ Hessian matrices
/ Normal distribution
/ Optimization
/ Parameter estimation
/ Polynomials
/ Theorems
2024
Please be aware that the book you have requested cannot be checked out. If you would like to checkout this book, you can reserve another copy
We have requested the book for you!
Your request is successful and it will be processed during the Library working hours. Please check the status of your request in My Requests.
Oops! Something went wrong.
Looks like we were not able to place your request. Kindly try again later.
Variance-Reduced Accelerated First-order Methods: Central Limit Theorems and Confidence Statements
Paper
Variance-Reduced Accelerated First-order Methods: Central Limit Theorems and Confidence Statements
2024
Request Book From Autostore
and Choose the Collection Method
Overview
In this paper, we study a stochastic strongly convex optimization problem and propose three classes of variable sample-size stochastic first-order methods including the standard stochastic gradient descent method, its accelerated variant, and the stochastic heavy ball method. In the iterates of each scheme, the unavailable exact gradients are approximated by averaging across an increasing batch size of sampled gradients. We prove that when the sample-size increases geometrically, the generated estimates converge in mean to the optimal solution at a geometric rate. Based on this result, we provide a unified framework to show that the rescaled estimation errors converge in distribution to a normal distribution, in which the covariance matrix depends on the Hessian matrix, covariance of the gradient noise, and the steplength. If the sample-size increases at a polynomial rate, we show that the estimation errors decay at the corresponding polynomial rate and establish the corresponding central limit theorems (CLTs). Finally, we provide an avenue to construct confidence regions for the optimal solution based on the established CLTs, and test the theoretic findings on a stochastic parameter estimation problem.
Publisher
Cornell University Library, arXiv.org
Subject
This website uses cookies to ensure you get the best experience on our website.