The alternating direction method of multipliers (ADMM) is widely used in solving structured convex optimization problems. Despite of its success in practice, the convergence of the standard ADMM for minimizing the sum of N (N ≥ 3) convex functions whose variables are linked by linear constraints, has remained unclear for a very long time. Recently, Chen et al.[4] provided a counterexample showing that the ADMM for N ≥ 3 may fail to converge without further conditions. Since the ADMM for N ≥ 3 has been very successful when applied to many problems arising from real practice, it is worth further investigating under what kind of sufficient conditions it can be guaranteed to converge. In this paper, we present such sufficient conditions that can guarantee the sublinear convergence rate for the ADMM for N ≥ 3. Specifically, we show that if one of the functions is convex (not necessarily strongly convex) and the other N − 1 functions are strongly convex, and the penalty parameter lies in a certain region, the ADMM converges with rate O(1/t) in a certain ergodic sense, and o(1/t) in a certain non-ergodic sense, where t denotes the number of iterations. As a by-product, we also provide a simple proof for the O(1/t) convergence rate of two-block ADMM in terms of both objective error and constraint violation, without assuming any condition on the penalty parameter and strong convexity on the functions.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.