Every local minimizer of a smooth constrained optimization problem satisfies the sequential approximate Karush-Kuhn-Tucker (AKKT) condition. This optimality condition is used to define the stopping criteria of many practical nonlinear programming algorithms. It is natural to ask for conditions on the constraints under which AKKT implies KKT. These conditions will be called strict constraint qualifications (SCQs). In this paper we define a cone-continuity property (CCP) that will be shown to be the weakest possible SCQ. Its relation to other constraint qualifications will also be clarified. In particular, it will be proved that CCP is strictly weaker than the constant positive generator constraint qualification.
A trio of genes might potentially distinguish doxorubicin-responsive from nonresponsive tumors, but further validation by a larger number of samples is still needed.
We present two new constraint qualifications (CQ) that are weaker than the recently introduced Relaxed Constant Positive Linear Dependence (RCPLD) constraint qualification. RCPLD is based on the assumption that many subsets of the gradients of the active constraints preserve positive linear dependence locally. A major open question was to identify the exact set of gradients whose properties had to be preserved locally and that would still work as a CQ. This is done in the first new constraint qualification, that we call Constant Rank of the Subspace Component (CRSC) CQ. This new CQ also preserves many of the good properties of RCPLD, like local stability and the validity of an error bound. We also introduce an even weaker CQ, called Constant Positive Generator (CPG), that can replace RCPLD in the analysis of the global convergence of algorithms. We close this work extending convergence results of algorithms belonging to all the main classes of nonlinear optimization methods: SQP, augmented Lagrangians, interior point algorithms, and inexact restoration. * This work was supported by PRONEX-Optimization (PRONEX-CNPq/FAPERJ E-26/171.510/2006-APQ1), Fapesp (Grants
Sequential optimality conditions have recently played an important role on the analysis of the global convergence of optimization algorithms towards first-order stationary points, justifying their stopping criteria. In this paper we introduce a sequential optimality condition that takes into account second-order information and that allows us to improve the global convergence assumptions of several second-order algorithms, which is our main goal. We also present a companion constraint qualification that is less stringent than previous assumptions associated to the convergence of second-order methods, like the joint condition Mangasarian-Fromovitz and Weak Constant Rank. Our condition is also weaker than the Constant Rank Constraint Qualification, which associated it to the convergence of second-order algorithms. This means that we can prove second-order global convergence of well stablished algorithms even when the set of Lagrange multipliers is unbounded, which overcomes a limitation of previous results based on MFCQ. We prove global convergence of well known variations of the augmented Lagrangian and Regularized SQP methods to second-order stationary points under this new weak constraint qualification.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.