Talks
Fall 2017
Regularized Nonlinear Acceleration
Tuesday, November 28th, 2017, 11:00 am–11:30 am
We describe a convergence acceleration technique for generic optimization problems. Our scheme computes estimates of the optimum from a nonlinear average of the iterates produced by any optimization method. The weights in this average are computed via a simple linear system, whose solution can be updated online. This acceleration scheme runs in parallel to the base algorithm, providing improved estimates of the solution on the fly, while the original optimization method is running. We apply this acceleration technique to stochastic algorithms such as SGD, SAGA, SVRG and Katyusha in different settings, and show significant performance gains.
Attachment | Size |
---|---|
Regularized Nonlinear Acceleration | 703.6 KB |