Loading…
NIPS 2013 has ended
Saturday, December 7 • 11:56am - 12:00pm
Non-strongly-convex smooth stochastic approximation with convergence rate O(1/n)

Sign up or log in to save this to your schedule, view media, leave feedback and see who's attending!

We consider the stochastic approximation problem where a convex function has to be minimized, given only the knowledge of unbiased estimates of its gradients at certain points, a framework which includes machine learning methods based on the minimization of the empirical risk. We focus on problems without strong convexity, for which all previously known algorithms achieve a convergence rate for function values of $O(1/\sqrt{n})$. We consider and analyze two algorithms that achieve a rate of $O(1/n)$ for classical supervised learning problems. For least-squares regression, we show that averaged stochastic gradient descent with constant step-size achieves the desired rate. For logistic regression, this is achieved by a simple novel stochastic gradient algorithm that (a) constructs successive local quadratic approximations of the loss functions, while (b) preserving the same running time complexity as stochastic gradient descent. For these algorithms, we provide a non-asymptotic analysis of the generalization error (in expectation, and also in high probability for least-squares), and run extensive experiments showing that they often outperform existing approaches.
None

Speakers
FB

Francis Bach

Francis Bach is a researcher in the Willow INRIA project-team, in the Computer Science Department of Ecole Normale Supérieure, Paris, France. He graduated from Ecole Polytechnique, Palaiseau, France, in 1997, and earned his PhD in 2005 from the Computer Science division at the University... Read More →


Saturday December 7, 2013 11:56am - 12:00pm PST
Harvey's Convention Center Floor, CC
  Spotlights
  • posterid Sat59
  • location Poster# Sat59