A note on solving nonlinear optimization problems in variable precision - Algorithmes Parallèles et Optimisation Accéder directement au contenu
Article Dans Une Revue Computational Optimization and Applications Année : 2020

A note on solving nonlinear optimization problems in variable precision

Résumé

This short note considers an efficient variant of the trust-region algorithm with dynamic accuracy proposed Carter (1993) and by Conn, Gould and Toint (2000) as a tool for very high-performance computing, an area where it is critical to allow multi-precision computations for keeping the energy dissipation under control. Numerical experiments are presented indicating that the use of the considered method can bring substantial savings in objective function's and gradient's evaluation "energy costs" by efficiently exploiting multi-precision computations. 1 Motivation and objectives Two recent evolutions in the field of scientific computing motivate the present note. The first is the growing importance of deep-learning methods for artificial intelligence, and the second is the acknowledgement by computer architects that new high-performance machines must be able to run the basic tools of deep learning very efficiently. Because the ubiquitous mathematical problem in deep learning is nonlinear nonconvex optimization, it is therefore of interest to consider how to solve this problem in ways that are as efficient as possible on new very powerful computers. As it turns out, one of the crucial aspects in designing such machines and the algorithms that they use is mastering energy dissipation. Given that this dissipation is approximately proportional to chip surface and that chip surface itself is approximately proportional to the square of the number of binary digits involved in the calculation [19, 31, 22, 26], being able to solve nonlinear optimization problems with as few digits as possible (while not loosing on final accuracy) is clearly of interest. This short note's sole purpose is to show that this is possible and that algorithms exist which achieve this goal and whose robustness significantly exceed simple minded approaches. The focus is on unconstrained nonconvex optimization, the most frequent case in deep learning applications. Since the cost of solving such problems is typically dominated by that of evaluating the objective function (and derivatives if possible), our aim is therefore to propose
Fichier principal
Vignette du fichier
1812.03467.pdf (229.15 Ko) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)
Loading...

Dates et versions

hal-02943237 , version 1 (18-09-2020)

Identifiants

Citer

Serge Gratton, Ph L Toint. A note on solving nonlinear optimization problems in variable precision. Computational Optimization and Applications, 2020, pp.1-15. ⟨hal-02943237⟩
65 Consultations
164 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More