Taming the Wild: A Unified Analysis of Hogwild!-Style Algorithms.

Journal: Advances in neural information processing systems
Published Date:

Abstract

Stochastic gradient descent (SGD) is a ubiquitous algorithm for a variety of machine learning problems. Researchers and industry have developed several techniques to optimize SGD's runtime performance, including asynchronous execution and reduced precision. Our main result is a martingale-based analysis that enables us to capture the rich noise models that may arise from such techniques. Specifically, we use our new analysis in three ways: (1) we derive convergence rates for the convex case (Hogwild!) with relaxed assumptions on the sparsity of the problem; (2) we analyze asynchronous SGD algorithms for non-convex matrix problems including matrix completion; and (3) we design and analyze an asynchronous SGD algorithm, called Buckwild!, that uses lower-precision arithmetic. We show experimentally that our algorithms run efficiently for a variety of problems on modern hardware.

Authors

  • Christopher De Sa
    Departments of Electrical Engineering and Computer Science, Stanford University.
  • Ce Zhang
    Stanford University.
  • Kunle Olukotun
    Stanford University.
  • Christopher RĂ©
    1Stanford University, Stanford, CA USA.

Keywords

No keywords available for this article.