4
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Stochastic Gradient Descent Escapes Saddle Points Efficiently

      Preprint
      , , , ,

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          This paper considers the perturbed stochastic gradient descent algorithm and shows that it finds \(\epsilon\)-second order stationary points (\(\left\|\nabla f(x)\right\|\leq \epsilon\) and \(\nabla^2 f(x) \succeq -\sqrt{\epsilon} \mathbf{I}\)) in \(\tilde{O}(d/\epsilon^4)\) iterations, giving the first result that has linear dependence on dimension for this setting. For the special case, where stochastic gradients are Lipschitz, the dependence on dimension reduces to polylogarithmic. In addition to giving new results, this paper also presents a simplified proof strategy that gives a shorter and more elegant proof of previously known results (Jin et al. 2017) on perturbed gradient descent algorithm.

          Related collections

          Most cited references2

          • Record: found
          • Abstract: not found
          • Article: not found

          Iterative hard thresholding for compressed sensing

            Bookmark
            • Record: found
            • Abstract: not found
            • Conference Proceedings: not found

            Finding approximate local minima faster than gradient descent

              Bookmark

              Author and article information

              Journal
              13 February 2019
              Article
              1902.04811
              470b3912-d5a0-476b-8316-b7306c6e8154

              http://arxiv.org/licenses/nonexclusive-distrib/1.0/

              History
              Custom metadata
              cs.LG math.OC stat.ML

              Numerical methods,Machine learning,Artificial intelligence
              Numerical methods, Machine learning, Artificial intelligence

              Comments

              Comment on this article