15th EUROPT Workshop on Advances in Continuous Optimization
Montréal, Canada, 12 — 14 July 2017
15th EUROPT Workshop on Advances in Continuous Optimization
Montréal, Canada, 12 — 14 July 2017
Nonlinear Optimization
Jul 12, 2017 03:30 PM – 05:10 PM
Location: PWC
Chaired by Florian Potra
4 Presentations

03:30 PM  03:55 PM
Numerical Observation of Maratos Effect with Multiple Precision Arithmetic
When we use the sequential quadratic programming (SQP) method to solve nonlinear programming problems (NLPs), the unit stepsize is required for superlinear convergence. However, we may suffer from the Maratos effect, which is a phenomenon in which the unit stepsize is rejected, even if an iteration point is sufficiently close to the optimal solution and the assumptions for fast convergence are satisfied. In this research, we observe this phenomenon numerically and seek some clues to overcome it efficiently. Especially, we utilize multiple precision arithmetic to observe it closely.

03:55 PM  04:20 PM
Adaptive matrix algebras in unconstrained optimization
In this communication we will introduce some recent techniques which involve structured matrix spaces in the reduction of time and space complexity of BFGStype minimization algorithms [1,2]. Some general results for the global convergence of algorithms for unconstrained optimization based on a BFGStype Hessian approximation scheme are introduced and it is shown how the constructibility of convergent algorithms suitable for large scale problems can be tackled using projections onto low complexity matrix algebras.
[1] C.Di Fiore, S.Fanelli, F.Lepore, P.Zellini, Matrix algebras in QuasiNewton
methods for unconstrained minimization, Numerische Mathematik, 94, (2003).[2] S.Cipolla, C.Di Fiore, F.Tudisco, P.Zellini, Adaptive matrix algebras in un
constrained minimization, Linear Algebra and its Application, 471, (2015). 
04:20 PM  04:45 PM
Efficient modifications on the parameters of Yabe and Takano’s conjugate gradient algorithm
Our main concern in this work is to make some modifications on the Yabe and Takano’s conjugate gradient (CG) algorithm [Comput. Optim. Appl. 28 (2004), pp. 203225] in order for receiving some appealing results in theory and practice. In fact, we propose an efficient adaptive updating formula for the parameters of Yabe and Takano CG algorithm leading to the well promising theoretical and numerical results. Global convergence property of the new proposed CG algorithm in DaiLiao family is established under standard assumptions on uniformly convex and general functions. Numerical performance of the new algorithm on some test problems from CUTEr collection shows the efficiency and effectiveness of the proposed method in practice.

04:45 PM  05:10 PM
A superquadratic variant of Newton's method
We present the first Qsuperquadratically convergent version of Newton's method for solving operator equations in Banach spaces that requires only one operator value and one inverse of the Frechet derivative per iteration. The Rorder of convergence is at least 2.4142. A semilocal analysis provides sufficient conditions for existence of a solution and convergence. The local analysis assumes that a solution exists and shows that the method converges from any starting point belonging to an explicitly defined neighbourhood of the solution called the ball of attraction.