Faster Coordinate Descent via Adaptive Importance Sampling
Coordinate descent methods employ random partial updates of decision variables in order to solve huge-scale convex optimization problems. In this work, we introduce new adaptive rules for the random selection of their updates. By adaptive, we mean that our selection rules are based on the dual residual or the primal-dual gap estimates and can change at each iteration. We theoretically characterize the performance of our selection rules and demonstrate improvements over the state-of-the-art, and extend our theory and algorithms to general convex objectives. Numerical evidence with hinge-loss support vector machines and Lasso confirm that the practice follows the theory.
284.pdf
openaccess
456.82 KB
Adobe PDF
d9bb22ee5c38fede23c0948c34631a9e
284-supp.pdf
openaccess
529.65 KB
Adobe PDF
26f4d756240833faa3f456907bff272d