Dmitriy Drusvyatskiy and I wrote a blog post about our recent paper *Stochastic subgradient method converges at the rate ** on weakly convex functions* on the ADSI blog, which you can view here.

# Tag: algorithms

# SMART: The Stochastic Monotone Aggregated Root-Finding Algorithm

Last night, I uploaded a new paper on the Stochastic Monotone Aggregated Root-Finding (SMART) algorithm. The algorithm excites me for a few reasons:

- SMART extends SAGA, SVRG, Finito, and SDCA, all of which solve problems like
to allow asynchronous parallel implementations, arbitrary block-coordinate updates, mini batching, and importance sampling.

- SMART replaces function gradients, , with black-boxes, , called
*operators*, and arrives at the root-finding problem:For SMART to converge, these operators need only satisfy a weak property, which we call the

*coherence condition*. - Because SMART works with operators, it generates some new algorithms for large-scale optimization problems like
where the functions are proximable and the maps are linear—these problems are hot in machine learning right now.

In the coming weeks, I’ll devote some blog posts to implementations of SMART on problems like logistic regression, support vector machines, collaborative filtering, feasibility problems, and more. In the meantime, check out the paper; comments are welcome.