JMLR 2015
An Asynchronous Parallel Stochastic Coordinate Descent Algorithm
Abstract
We describe an asynchronous parallel stochastic coordinate descent algorithm for minimizing smooth unconstrained or separably constrained functions. The method achieves a linear convergence rate on functions that satisfy an essential strong convexity property and a sublinear rate ($1/K$) on general convex functions. Near-linear speedup on a multicore system can be expected if the number of processors is $O(n^{1/2})$ in unconstrained optimization and $O(n^{1/4})$ in the separable- constrained case, where $n$ is the number of variables. We describe results from implementation on 40-core processors. [abs] [ pdf ][ bib ] © JMLR 2015. ( edit, beta )
Authors
Keywords
No keywords are indexed for this paper.
Context
- Venue
- Journal of Machine Learning Research
- Archive span
- 2000-2026
- Indexed papers
- 4180
- Paper id
- 945177337552219388