A derivative-free two level random search method for unconstrained optimization / Neculai Andrei.
2021
QA402.5
Linked e-resources
Linked Resource
Online Access
Concurrent users
Unlimited
Authorized users
Authorized users
Document Delivery Supplied
Can lend chapters, not whole ebooks
Details
Title
A derivative-free two level random search method for unconstrained optimization / Neculai Andrei.
Author
Andrei, Neculai.
ISBN
9783030685171 (electronic bk.)
3030685179 (electronic bk.)
9783030685188 (print)
3030685187
3030685160
9783030685164
3030685179 (electronic bk.)
9783030685188 (print)
3030685187
3030685160
9783030685164
Publication Details
Cham, Switzerland : Springer, 2021.
Language
English
Description
1 online resource
Item Number
10.1007/978-3-030-68517-1 doi
Call Number
QA402.5
Dewey Decimal Classification
519.6
Summary
The book is intended for graduate students and researchers in mathematics, computer science, and operational research. The book presents a new derivative-free optimization method/algorithm based on randomly generated trial points in specified domains and where the best ones are selected at each iteration by using a number of rules. This method is different from many other well established methods presented in the literature and proves to be competitive for solving many unconstrained optimization problems with different structures and complexities, with a relative large number of variables. Intensive numerical experiments with 140 unconstrained optimization problems, with up to 500 variables, have shown that this approach is efficient and robust. Structured into 4 chapters, Chapter 1 is introductory. Chapter 2 is dedicated to presenting a two level derivative-free random search method for unconstrained optimization. It is assumed that the minimizing function is continuous, lower bounded and its minimum value is known. Chapter 3 proves the convergence of the algorithm. In Chapter 4, the numerical performances of the algorithm are shown for solving 140 unconstrained optimization problems, out of which 16 are real applications. This shows that the optimization process has two phases: the reduction phase and the stalling one. Finally, the performances of the algorithm for solving a number of 30 large-scale unconstrained optimization problems up to 500 variables are presented. These numerical results show that this approach based on the two level random search method for unconstrained optimization is able to solve a large diversity of problems with different structures and complexities. There are a number of open problems which refer to the following aspects: the selection of the number of trial or the number of the local trial points, the selection of the bounds of the domains where the trial points and the local trial points are randomly generated and a criterion for initiating the line search.
Bibliography, etc. Note
Includes bibliographical references.
Access Note
Access limited to authorized users.
Digital File Characteristics
text file
PDF
Series
SpringerBriefs in optimization.
Available in Other Form
Derivative-free two level random search method for unconstrained optimization.
Linked Resources
Online Access
Record Appears in
Online Resources > Ebooks
All Resources
All Resources
Table of Contents
1. Introduction
2. A Derivative-free Two Level Random Search Method for Unconstrained Optimization
3. Convergence of the Algorithm
4. Numerical Results
5. Conclusions
Annex A. List of Applications
Annex B. List of Test Functions
Annex C. Detailed Results for 30 Large-Scale Problems
Annex D. Detailed Results for 140 Problems.
2. A Derivative-free Two Level Random Search Method for Unconstrained Optimization
3. Convergence of the Algorithm
4. Numerical Results
5. Conclusions
Annex A. List of Applications
Annex B. List of Test Functions
Annex C. Detailed Results for 30 Large-Scale Problems
Annex D. Detailed Results for 140 Problems.