A new method of moving asymptotes for large-scale minimization subject to linear equality constraints is discussed. In this method, linear equality constraints are deleted with null space technique and the descending direction is obtained by solving a convex separable subproblem of moving asymptotes in each iteration. New rules for controlling the asymptotes parameters are designed and the global convergence of the method under some reasonable conditions is established and proved. The numerical results show that the new method may be capable of processing some large scale problems.
The new trust region subproblem with the conic model was proposed in 2005, and was divided into three different cases. The first two cases can be converted into a quadratic model or a convex problem with quadratic constraints, while the third one is a nonconvex problem. In this paper, first we analyze the nonconvex problem, and reduce it to two convex problems. Then we discuss some dual properties of these problems and give an algorithm for solving them. At last, we present an algorithm for solving the new trust region subproblem with the conic model and report some numerical examples to illustrate the efficiency of the algorithm.
A new limited memory symmetric rank one algorithm is proposed. It combines a modified self-scaled symmetric rank one (SSR1) update with the limited memory and nonmonotone line search technique. In this algorithm, the descent search direction is generated by inverse limited memory SSR1 update, thus simplifying the computation. Numerical comparison of the algorithm and the famous limited memory BFGS algorithm is given. Comparison results indicate that the new algorithm can process a kind of large-scale unconstrained optimization problems.