In view of solving convex optimization problems with noisy gradient input, we analyze the asymptotic behavior of gradient-like flows under stochastic disturbances. Specifically, we focus on the widely studied class of mirror descent schemes for convex programs with compact feasible regions, and we examine the dynamics' convergence and concentration properties in the presence of noise. In the vanishing noise limit, we show that the dynamics converge to the solution set of the underlying problem (a.s.). Otherwise, when the noise is persistent, we show that the dynamics are concentrated around interior solutions in the long run, and they converge to boundary solutions that are sufficiently "sharp". Finally, we show that a suitably rectified variant of the method converges irrespective of the magnitude of the noise (or the structure of the underlying convex program), and we derive an explicit estimate for its rate of convergence.
We consider a co-evolutionary model of social coordination and network formation where agents may decide on an action in a 2 × 2-coordination game and on whom to establish costly links to. We find that a payoff dominant convention is selected for a wider parameter range when agents may only support a limited number of links as compared to a scenario where agents are not constrained in their linking choice. The main reason behind this result is that constrained interactions create a tradeoff between the interactions an agent has and those he would rather have. Further, we discuss convex linking costs and provide sufficient conditions for the payoff dominant convention to be selected in m × m coordination games.
We consider a model of stochastic evolution under general noisy best-response protocols, allowing the probabilities of suboptimal choices to depend on their payoff consequences. Our analysis focuses on behavior in the small noise double limit: we first take the noise level in agents' decisions to zero, and then take the population size to infinity. We show that in this double limit, escape from and transitions between equilibria can be described in terms of solutions to continuous optimal control problems. These are used in turn to characterize the asymptotics of the stationary distribution, and so to determine the stochastically stable states. We use these results to perform a complete analysis of evolution in threestrategy coordination games that satisfy the marginal bandwagon property and that have an interior equilibrium, with agents following the logit choice rule. This article merges two working papers: "Large deviations and stochastic stability in the small noise double limit, I: Theory" and "Large deviations and stochastic stability in the small noise double limit, II: The logit model." We thank a co-editor and a number of anonymous referees for valuable comments, and Daniel Liberzon for helpful discussions.
Projection-free optimization via different variants of the Frank–Wolfe method has become one of the cornerstones of large scale optimization for machine learning and computational statistics. Numerous applications within these fields involve the minimization of functions with self-concordance like properties. Such generalized self-concordant functions do not necessarily feature a Lipschitz continuous gradient, nor are they strongly convex, making them a challenging class of functions for first-order methods. Indeed, in a number of applications, such as inverse covariance estimation or distance-weighted discrimination problems in binary classification, the loss is given by a generalized self-concordant function having potentially unbounded curvature. For such problems projection-free minimization methods have no theoretical convergence guarantee. This paper closes this apparent gap in the literature by developing provably convergent Frank–Wolfe algorithms with standard $$\mathcal {O}(1/k)$$ O ( 1 / k ) convergence rate guarantees. Based on these new insights, we show how these sublinearly convergent methods can be accelerated to yield linearly convergent projection-free methods, by either relying on the availability of a local liner minimization oracle, or a suitable modification of the away-step Frank–Wolfe method.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
hi@scite.ai
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.