Optimizing radial basis functions by D.C. programming and its use in direct search for global derivative-free optimization

In this paper we address the global optimization of functions subject to bound and linear constraints without using derivatives of the objective function. We investigate the use of derivative-free models based on radial basis functions (RBFs) in the search step of direct-search methods of directional type. We also study the application of algorithms based on … Read more

Dynamic Evolution for Risk-Neutral Densities

Option price data is often used to infer risk-neutral densities for future prices of an underlying asset. Given the prices of a set of options on the same underlying asset with different strikes and maturities, we propose a nonparametric approach for estimating the evolution of the risk-neutral density in time. Our method uses bicubic splines … Read more

Incorporating Minimum Frobenius Norm Models in Direct Search

The goal of this paper is to show that the use of minimum Frobenius norm quadratic models can improve the performance of direct-search methods. The approach taken here is to maintain the structure of directional direct-search methods, organized around a search and a poll step, and to use the set of previously evaluated points generated … Read more

A globally convergent primal-dual interior-point filter method for nonlinear programming: new filter optimality measures and computational results

In this paper we modify the original primal-dual interior-point filter method proposed in [18] for the solution of nonlinear programming problems. We introduce two new optimality filter entries based on the objective function, and thus better suited for the purposes of minimization, and propose conditions for using inexact Hessians. We show that the global convergence … Read more

PSwarm: A Hybrid Solver for Linearly Constrained Global Derivative-Free Optimization

PSwarm was developed originally for the global optimization of functions without derivatives and where the variables are within upper and lower bounds. The underlying algorithm used is a pattern search method, more specifically a coordinate search method, which guarantees convergence to stationary points from arbitrary starting points. In the (optional) search step of coordinate search, … Read more

Implicitely and Densely Discrete Black-Box Optimization Problems

This paper addresses derivative-free optimization problems where the variables lie implicitly in an unknown discrete closed set. The evaluation of the objective function follows a projection onto the discrete set, which is assumed dense rather than sparse. Such a mathematical setting is a rough representation of what is common in many real-life applications where, despite … Read more

Using Simplex Gradients of Nonsmooth Functions in Direct Search Methods

It has been shown recently that the efficiency of direct search methods that use opportunistic polling in positive spanning directions can be improved significantly by reordering the poll directions according to descent indicators built from simplex gradients. The purpose of this paper is twofold. First, we analyze the properties of simplex gradients of nonsmooth functions … Read more

Global Convergence of General Derivative-Free Trust-Region Algorithms to First and Second Order Critical Points

In this paper we prove global convergence for first and second-order stationarity points of a class of derivative-free trust-region methods for unconstrained optimization. These methods are based on the sequential minimization of linear or quadratic models built from evaluating the objective function at sample sets. The derivative-free models are required to satisfy Taylor-type bounds but, … Read more

A Particle Swarm Pattern Search Method for Bound Constrained Nonlinear Optimization

In this paper we develop, analyze, and test a new algorithm for the global minimization of a function subject to simple bounds without the use of derivatives. The underlying algorithm is a pattern search method, more specifically a coordinate search method, which guarantees convergence to stationary points from arbitrary starting points. In the optional search … Read more

Geometry of Sample Sets in Derivative Free Optimization. Part II: Polynomial Regression and Underdetermined Interpolation

In the recent years, there has been a considerable amount of work in the development of numerical methods for derivative free optimization problems. Some of this work relies on the management of the geometry of sets of sampling points for function evaluation and model building. In this paper, we continue the work developed in [Conn, … Read more