Digitala Vetenskapliga Arkivet

Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Approaches to accelerate methods for solving systems of equations arising in nonlinear optimization
KTH, School of Engineering Sciences (SCI), Mathematics (Dept.), Optimization and Systems Theory.ORCID iD: 0000-0003-1764-5449
2020 (English)Doctoral thesis, comprehensive summary (Other academic)
Abstract [en]

Methods for solving nonlinear optimization problems typically involve solving systems of equations. This thesis concerns approaches for accelerating some of those methods. In our setting, accelerating involves finding a trade-off between the computational cost of an iteration and the quality of the computed search direction. We design approaches for which theoretical results in ideal settings are derived. We also investigate the practical performance of the approaches within and beyond the boundaries of the theoretical frameworks with numerical simulations.

Paper A concerns methods for solving strictly convex unconstrained quadratic optimization problems. This is equivalent to solving systems of linear equations where the matrices are symmetric positive definite. The main focus is exact linesearch limited-memory quasi-Newton methods which generate search directions parallel to those of the method of preconditioned conjugate gradients. We give a class of limited-memory quasi-Newton methods. In addition, we provide a dynamic framework for the construction of these methods. The methods are meant to be particularly useful for solving sequences of related systems of linear equations. Such sequences typically arise as methods for solving systems of nonlinear equations converge.

Paper B deals with solving systems of nonlinear equations that arise in interior-point methods for bound-constrained nonlinear programming. Application of Newton's method gives sequences of systems of linear equations. We propose partial and full approximate solutions to the Newton systems. The partial approximate solutions are computationally inexpensive, whereas each full approximate solution typically requires a reduced Newton system to be solved. In addition, we suggest two Newton-like approaches, which are based upon the proposed partial approximate solutions, for solving the systems of nonlinear equations.

Paper C is focused on interior-point methods for quadratic programming. We propose a structured modified Newton approach to solve the systems of nonlinear equations that arise. The modified Jacobians are composed of a previous Jacobian, plus one low-rank update matrix per succeeding iteration. For a given rank restriction, we construct a low-rank update matrix such that the modified Jacobian becomes closest to the current Jacobian, in both 2-norm and Frobenious norm. The approach is structured in the sense that it preserves the nonzero pattern of the Jacobian.

The approaches suggested in Paper B and Paper C are motivated by asymptotic results in ideal theoretical frameworks. In particular, it is shown that the approaches become increasingly accurate as primal-dual interior-point methods converge. A demonstration of the practical performance is given by numerical results. The results indicate the performance and limitations of the approaches suggested.

We envisage that the approaches of Paper A, Paper B and Paper C can be useful in parallel, or in combination, with existing solvers in order to accelerate methods.

Paper D is more pedagogical in nature. We give a derivation of the method of conjugate gradients in an optimization framework. The result itself is well known but the derivation has, to the best of our knowledge, not been presented before.

Abstract [sv]

Metoder för att lösa ickelinjära optimeringsproblem involverar vanligtvis lösning av ekvationssystem. Denna avhandling handlar om strategier för att accelerera några av dessa metoder. I vårt ramverk innebär acceleration att hitta en avvägning mellan beräkningskostnaden för en iteration och kvaliteten av den beräknade sökriktningen. Vi utformar strategier för vilka teoretiska resultat i ideala omgivningar härleds. Vi undersöker också strategiernas praktiska prestanda inom och utanför gränserna för de teoretiska omgivningarna med numeriska simuleringar.

Artikel A behandlar metoder för att lösa strikt konvexa kvadratiska optimeringsproblem utan bivillkor. Detta motsvarar att lösa linjära ekvationssystem där matriserna är symmetriska och positivt definita. Huvudfokus är kvasi-Newtonmetoder, med exakt linjesökning och begränsat minne, som genererar sökriktningar parallella med de i konjugerade gradientmetoden. Vi ger en klass av kvasi-Newtonmetoder med begränsat minne. Dessutom föreslår vi ett dynamiskt ramverk för att konstruera dessa metoder. Metoderna är avsedda att vara särskilt användbara för att lösa sekvenser av relaterade linjära ekvationssystem. Sådana sekvenser förekommer vanligtvis då metoder för att lösa ickelinjära ekvationssystem konvergerar.

Artikel B handlar om att lösa ickelinjära ekvationssystem som uppstår i inre-punktsmetoder för ickelinjär optimering med begränsade variabler. Till-ämpning av Newtons metod ger sekvenser av linjära ekvationssystem. Vi föreslår partiella och fullständiga approximativa lösningar till Newton-systemen. De partiella approximativa lösningarna är beräkningsmässigt billiga, medan varje fullständig approximativ lösning vanligtvis kräver att ett reducerat Newton-system löses. Dessutom föreslår vi två Newton-liknande strategier, som baseras på de föreslagna partiella approximativa lösningarna, för att lösa de ickelinjära ekvationssystemen.

Artikel C berör inre-punktsmetoder för kvadratisk optimering med linjära olikhetsbivillkor. Vi föreslår en strukturerad modifierad Newton-strategi för att lösa de ickelinjära ekvationssystemen som uppstår. De modifierade Jakobianerna är sammansatta av en tidigare Jakobian, samt en uppdateringsmatris av låg rang per efterföljande iteration. För en given rangbegränsning konstruerar vi en uppdateringsmatris av låg rang sådan att den modifierade Jakobianen blir närmast den nuvarande Jakobianen, i både 2-norm och Frobenious-norm. Strategin är strukturerad i avseendet att den bevarar Jakobianens gleshetsstruktur.

Strategierna som föreslås i Artikel B och Artikel C motiveras med asymptotiska resultat i ideala teoretiska omgivningar. Särskilt visas det att strategierna blir alltmer exakta då primal-duala inre-punktsmetoder konvergerar. En undersökning av deras praktiska prestanda ges genom numeriska resultat. Resultaten illustrerar de föreslagna strategiernas praktiska prestanda samt deras begränsningar.

Vi föreställer oss att strategierna i Artikel A, Artikel B och Artikel C kan vara användbara parallellt, eller i kombination, med befintliga lösare för att accelerera metoder.

Artikel D är mer pedagogisk i sin natur. Vi ger en härledning av konjugerade gradientsmetoden i ett optimeringsramverk. Resultatet i sig är välkänt men härledningen har, såvitt vi vet, inte presenterats tidigare.

Place, publisher, year, edition, pages
Stockholm, Sweden: KTH Royal Institute of Technology, 2020. , p. 24
Series
TRITA-SCI-FOU ; 2020:37
Keywords [en]
Nonlinear optimization, mathematical programming, interior-point methods, approximate solutions to systems of linear equations, method of conjugate gradients, quasi-Newton methods, modified Newton methods
Keywords [sv]
Ickelinjär optimering, matematisk programmering, inre-punktsmetoder, approximativa lösningar till linjära ekvationssystem, konjugerade gradientmetoden, kvasi-Newtonmetoder, modifierade Newtonmetoder.
National Category
Computational Mathematics
Research subject
Applied and Computational Mathematics, Optimization and Systems Theory
Identifiers
URN: urn:nbn:se:kth:diva-286057ISBN: 978-91-7873-713-0 (print)OAI: oai:DiVA.org:kth-286057DiVA, id: diva2:1502235
Public defence
2020-12-18, via Zoom, https://kth-se.zoom.us/j/63180134076, Stockholm, 10:00 (English)
Opponent
Supervisors
Funder
Swedish Research Council, 621-2014-4772Available from: 2020-11-19 Created: 2020-11-19 Last updated: 2022-06-25Bibliographically approved
List of papers
1. Exact linesearch limited-memory quasi-Newton methods for minimizing a quadratic function
Open this publication in new window or tab >>Exact linesearch limited-memory quasi-Newton methods for minimizing a quadratic function
(English)Manuscript (preprint) (Other academic)
Abstract [en]

The main focus in this paper is exact linesearch methods for minimizing a quadratic function whose Hessian is positive definite. We give a class of limited-memory quasi-Newton Hessian approximations which generate search directions parallel to those of the method of preconditioned conjugate gradients, and hence give finite termination on quadratic optimization problems in exact arithmetic. With the framework of reduced-Hessians this class  provides a dynamical framework for the construction of limited-memory quasi-Newton methods. We give an indication of the performance of the methods within this framework by showing numerical simulations on sequences of related systems of linear equations, which originate from the CUTEst test collection. In addition, we give a compact representation of the Hessian approximations in the full Broyden class for the general unconstrained optimization problem. This representation consists of explicit matrices and gradients only as vector components.

Keywords
method of conjugate gradients, quasi-Newton method, unconstrained quadratic program, limited-memory method, exact linesearch method
National Category
Computational Mathematics
Research subject
Applied and Computational Mathematics, Optimization and Systems Theory
Identifiers
urn:nbn:se:kth:diva-286030 (URN)
Funder
Swedish Research Council, 621-2014-4772
Note

QC 20201201

Available from: 2020-11-18 Created: 2020-11-18 Last updated: 2022-06-25Bibliographically approved
2. Approximate solution of system of equations arising in interior-point methods for bound-constrained optimization
Open this publication in new window or tab >>Approximate solution of system of equations arising in interior-point methods for bound-constrained optimization
(English)Manuscript (preprint) (Other academic)
Abstract [en]

The focus in this paper is interior-point methods for bound-constrained nonlinear optimization, where the system of nonlinear equations that arise are solved with Newton's method. There is a trade-off between solving Newton systems directly, which give high quality solutions, and solving many approximate Newton systems which are computationally less expensive but give lower quality solutions. We propose partial and full approximate solutions to the Newton systems. The specific approximate solution depends on estimates of the active and inactive constraints at the solution. These sets are at each iteration estimated by basic heuristics. The partial approximate solutions are computationally inexpensive, whereas a system of linear equations needs to be solved for the full approximate solution. The size of the system is determined by the estimate of the inactive constraints at the solution. In addition, we motivate and suggest two Newton-like approaches which are based on an intermediate step that consists of the partial approximate solutions. The theoretical setting is introduced and asymptotic error bounds are given. We also give numerical results to investigate the performance of the approximate solutions within and beyond the theoretical framework. 

Keywords
interior-point methods, bound-constrained optimization, approximate solution of system of linear equations, Newton-like approaches
National Category
Computational Mathematics
Research subject
Applied and Computational Mathematics, Optimization and Systems Theory
Identifiers
urn:nbn:se:kth:diva-286032 (URN)
Funder
Swedish Research Council, 621-2014-4772
Note

QC 20201201

Available from: 2020-11-18 Created: 2020-11-18 Last updated: 2022-06-25Bibliographically approved
3. A structured modified Newton approach for solving systems of nonlinear equations arising in interior-point methods for quadratic programming
Open this publication in new window or tab >>A structured modified Newton approach for solving systems of nonlinear equations arising in interior-point methods for quadratic programming
(English)Manuscript (preprint) (Other academic)
Abstract [en]

The focus in this work is interior-point methods for quadratic optimization problems with linear inequality constraints where the system of nonlinear equations that arise are solved with Newton-like methods. In particular, the concern is the system of linear equations to be solved at each iteration. Newton systems give high quality solutions but there is an interest in designing modified Newton systems which are computationally less expensive but normally at the expense of lower quality solutions. We propose a structured modified Newton approach where the Jacobian of each Newton system is approximated by a Jacobian at a previous point plus a sequence of low-rank update matrices. The updates are such that the Jacobian approximation preserves its sparsity structure and in consequence may be viewed as a Jacobian evaluated at a different point. We introduce the theoretical setting, motivate the approach by theoretical results applicable in the asymptotic region and give numerical results for a set of convex quadratic programs. In an attempt to improve performance we also motivate and construct two heuristics to be added to the method.

Keywords
interior-point methods, modified Newton methods, quasi-Newton methods, low-rank updates
National Category
Computational Mathematics
Research subject
Applied and Computational Mathematics, Optimization and Systems Theory
Identifiers
urn:nbn:se:kth:diva-286034 (URN)
Note

QC 20201201

Available from: 2020-11-18 Created: 2020-11-18 Last updated: 2022-06-25Bibliographically approved
4. An optimization derivation of the method of conjugate gradients
Open this publication in new window or tab >>An optimization derivation of the method of conjugate gradients
(English)Manuscript (preprint) (Other academic)
Abstract [en]

We give a derivation of the method of conjugate gradients based on the requirement that each iterate minimizes a strictly convex quadratic on the space spanned by the previously observed gradients. Rather than verifying that the search direction has the correct properties, we show that generation of such iterates is equivalent to generation of orthogonal gradients which gives the description of the direction and the step length. Our approach gives a straightforward way to see that the searchdirection of the method of conjugate gradients is a negative scalartimes the gradient of minimum Euclidean norm on the subspace generated so far.

Keywords
method of conjugate gradients, orthogonal gradients, gradient of minimum Euclidean norm.
National Category
Computational Mathematics
Research subject
Applied and Computational Mathematics, Optimization and Systems Theory
Identifiers
urn:nbn:se:kth:diva-286036 (URN)
Funder
Swedish Research Council, 621-2014-4772
Note

QC 20201201

Available from: 2020-11-18 Created: 2020-11-18 Last updated: 2022-06-25Bibliographically approved

Open Access in DiVA

fulltext(519 kB)686 downloads
File information
File name FULLTEXT01.pdfFile size 519 kBChecksum SHA-512
ca2dacd61f8fa34296c1cadb8a41a30d1c1e8b7cdc9da59e9ee7e12f03f2f6ffdc2f042b2837c9d789bcf19468a44bd746c4a6e9b5f5255e3a8b80dfbdadc321
Type fulltextMimetype application/pdf

Search in DiVA

By author/editor
Ek, David
By organisation
Optimization and Systems Theory
Computational Mathematics

Search outside of DiVA

GoogleGoogle Scholar
Total: 686 downloads
The number of downloads is the sum of all downloads of full texts. It may include eg previous versions that are now no longer available

isbn
urn-nbn

Altmetric score

isbn
urn-nbn
Total: 2016 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf