This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

\coltauthor\Name

Cedar Site Bai \Emailbai123@purdue.edu
\addrDepartment of Computer Science
Purdue University and \NameBrian Bullins \Emailbbullins@purdue.edu
\addrDepartment of Computer Science
Purdue University

Faster Acceleration for Steepest Descent

Abstract

Recent advances (Sherman2017Area; sidford2018coordinate; cohen2021relative) have overcome the fundamental barrier of dimension dependence in the iteration complexity of solving \ell_{\infty} regression with first-order methods. Yet it remains unclear to what extent such acceleration can be achieved for general p\ell_{p} smooth functions. In this paper, we propose a new accelerated first-order method for convex optimization under non-Euclidean smoothness assumptions. In contrast to standard acceleration techniques, our approach uses primal-dual iterate sequences taken with respect to differing norms, which are then coupled using an implicitly determined interpolation parameter. For p\ell_{p} norm smooth problems in dd dimensions, our method provides an iteration complexity improvement of up to O(d12p)O(d^{1-\frac{2}{p}}) in terms of calls to a first-order oracle, thereby allowing us to circumvent long-standing barriers in accelerated non-Euclidean steepest descent.

keywords:
First-order acceleration, convex optimization, non-Euclidean smoothness, steepest descent

1 Introduction

Large-scale optimization tasks are a central part of modern machine learning, and many of the algorithms that find success in training these models, such as SGD