| | """Abstraction for optimizers. |
| | |
| | It is sufficient that one re-implements the base estimator. |
| | """ |
| |
|
| | import numbers |
| | import warnings |
| |
|
| | try: |
| | from collections.abc import Iterable |
| | except ImportError: |
| | from collections.abc import Iterable |
| |
|
| | from ..callbacks import VerboseCallback, check_callback |
| | from ..utils import eval_callbacks |
| | from .optimizer import Optimizer |
| |
|
| |
|
| | def base_minimize( |
| | func, |
| | dimensions, |
| | base_estimator, |
| | n_calls=100, |
| | n_random_starts=None, |
| | n_initial_points=10, |
| | initial_point_generator="random", |
| | acq_func="EI", |
| | acq_optimizer="lbfgs", |
| | x0=None, |
| | y0=None, |
| | random_state=None, |
| | verbose=False, |
| | callback=None, |
| | n_points=10000, |
| | n_restarts_optimizer=5, |
| | xi=0.01, |
| | kappa=1.96, |
| | n_jobs=1, |
| | model_queue_size=None, |
| | space_constraint=None, |
| | ): |
| | """Base optimizer class. |
| | |
| | Parameters |
| | ---------- |
| | func : callable |
| | Function to minimize. Should take a single list of parameters |
| | and return the objective value. |
| | |
| | If you have a search-space where all dimensions have names, |
| | then you can use :func:`skopt.utils.use_named_args` as a decorator |
| | on your objective function, in order to call it directly |
| | with the named arguments. See `use_named_args` for an example. |
| | |
| | dimensions : list, shape (n_dims,) |
| | List of search space dimensions. |
| | Each search dimension can be defined either as |
| | |
| | - a `(lower_bound, upper_bound)` tuple (for `Real` or `Integer` |
| | dimensions), |
| | - a `(lower_bound, upper_bound, "prior")` tuple (for `Real` |
| | dimensions), |
| | - as a list of categories (for `Categorical` dimensions), or |
| | - an instance of a `Dimension` object (`Real`, `Integer` or |
| | `Categorical`). |
| | |
| | .. note:: The upper and lower bounds are inclusive for `Integer` |
| | dimensions. |
| | |
| | base_estimator : sklearn regressor |
| | Should inherit from `sklearn.base.RegressorMixin`. |
| | In addition, should have an optional `return_std` argument, |
| | which returns `std(Y | x)` along with `E[Y | x]`. |
| | |
| | n_calls : int, default: 100 |
| | Maximum number of calls to `func`. An objective function will |
| | always be evaluated this number of times; Various options to |
| | supply initialization points do not affect this value. |
| | |
| | n_random_starts : int, default: None |
| | Number of evaluations of `func` with random points before |
| | approximating it with `base_estimator`. |
| | |
| | .. deprecated:: 0.8 |
| | use `n_initial_points` instead. |
| | |
| | n_initial_points : int, default: 10 |
| | Number of evaluations of `func` with initialization points |
| | before approximating it with `base_estimator`. Initial point |
| | generator can be changed by setting `initial_point_generator`. |
| | |
| | initial_point_generator : str, InitialPointGenerator instance, \ |
| | default: `"random"` |
| | Sets a initial points generator. Can be either |
| | |
| | - `"random"` for uniform random numbers, |
| | - `"sobol"` for a Sobol' sequence, |
| | - `"halton"` for a Halton sequence, |
| | - `"hammersly"` for a Hammersly sequence, |
| | - `"lhs"` for a latin hypercube sequence, |
| | - `"grid"` for a uniform grid sequence |
| | |
| | acq_func : string, default: `"EI"` |
| | Function to minimize over the posterior distribution. Can be either |
| | |
| | - `"LCB"` for lower confidence bound, |
| | - `"EI"` for negative expected improvement, |
| | - `"PI"` for negative probability of improvement. |
| | - `"EIps"` for negated expected improvement per second to take into |
| | account the function compute time. Then, the objective function is |
| | assumed to return two values, the first being the objective value and |
| | the second being the time taken in seconds. |
| | - `"PIps"` for negated probability of improvement per second. The |
| | return type of the objective function is assumed to be similar to |
| | that of `"EIps"` |
| | |
| | acq_optimizer : string, `"sampling"` or `"lbfgs"`, default: `"lbfgs"` |
| | Method to minimize the acquisition function. The fit model |
| | is updated with the optimal value obtained by optimizing `acq_func` |
| | with `acq_optimizer`. |
| | |
| | - If set to `"sampling"`, then `acq_func` is optimized by computing |
| | `acq_func` at `n_points` randomly sampled points and the smallest |
| | value found is used. |
| | - If set to `"lbfgs"`, then |
| | |
| | - The `n_restarts_optimizer` no. of points which the acquisition |
| | function is least are taken as start points. |
| | - `"lbfgs"` is run for 20 iterations with these points as initial |
| | points to find local minima. |
| | - The optimal of these local minima is used to update the prior. |
| | |
| | x0 : list, list of lists or `None` |
| | Initial input points. |
| | |
| | - If it is a list of lists, use it as a list of input points. If no |
| | corresponding outputs `y0` are supplied, then len(x0) of total |
| | calls to the objective function will be spent evaluating the points |
| | in `x0`. If the corresponding outputs are provided, then they will |
| | be used together with evaluated points during a run of the algorithm |
| | to construct a surrogate. |
| | - If it is a list, use it as a single initial input point. The |
| | algorithm will spend 1 call to evaluate the initial point, if the |
| | outputs are not provided. |
| | - If it is `None`, no initial input points are used. |
| | |
| | y0 : list, scalar or `None` |
| | Objective values at initial input points. |
| | |
| | - If it is a list, then it corresponds to evaluations of the function |
| | at each element of `x0` : the i-th element of `y0` corresponds |
| | to the function evaluated at the i-th element of `x0`. |
| | - If it is a scalar, then it corresponds to the evaluation of the |
| | function at `x0`. |
| | - If it is None and `x0` is provided, then the function is evaluated |
| | at each element of `x0`. |
| | |
| | random_state : int, RandomState instance, or None (default) |
| | Set random state to something other than None for reproducible |
| | results. |
| | |
| | verbose : boolean, default: False |
| | Control the verbosity. It is advised to set the verbosity to True |
| | for long optimization runs. |
| | |
| | callback : callable, list of callables, optional |
| | If callable then `callback(res)` is called after each call to `func`. |
| | If list of callables, then each callable in the list is called. |
| | |
| | n_points : int, default: 10000 |
| | If `acq_optimizer` is set to `"sampling"`, then `acq_func` is |
| | optimized by computing `acq_func` at `n_points` randomly sampled |
| | points. |
| | |
| | n_restarts_optimizer : int, default: 5 |
| | The number of restarts of the optimizer when `acq_optimizer` |
| | is `"lbfgs"`. |
| | |
| | xi : float, default: 0.01 |
| | Controls how much improvement one wants over the previous best |
| | values. Used when the acquisition is either `"EI"` or `"PI"`. |
| | |
| | kappa : float, default: 1.96 |
| | Controls how much of the variance in the predicted values should be |
| | taken into account. If set to be very high, then we are favouring |
| | exploration over exploitation and vice versa. |
| | Used when the acquisition is `"LCB"`. |
| | |
| | n_jobs : int, default: 1 |
| | Number of cores to run in parallel while running the lbfgs |
| | optimizations over the acquisition function and given to |
| | the base_estimator. Valid only when |
| | `acq_optimizer` is set to "lbfgs". or when the base_estimator |
| | supports n_jobs as parameter and was given as string. |
| | Defaults to 1 core. If `n_jobs=-1`, then number of jobs is set |
| | to number of cores. |
| | |
| | model_queue_size : int or None, default: None |
| | Keeps list of models only as long as the argument given. In the |
| | case of None, the list has no capped length. |
| | |
| | space_constraint : callable or None, default: None |
| | Constraint function. Should take a single list of parameters |
| | (i.e. a point in space) and return True if the point satisfies |
| | the constraints. |
| | If None, the space is not conditionally constrained. |
| | |
| | Returns |
| | ------- |
| | res : `OptimizeResult`, scipy object |
| | The optimization result returned as a OptimizeResult object. |
| | Important attributes are: |
| | |
| | - `x` [list]: location of the minimum. |
| | - `fun` [float]: function value at the minimum. |
| | - `models`: surrogate models used for each iteration. |
| | - `x_iters` [list of lists]: location of function evaluation for each |
| | iteration. |
| | - `func_vals` [array]: function value for each iteration. |
| | - `space` [Space]: the optimization space. |
| | - `specs` [dict]`: the call specifications. |
| | - `rng` [RandomState instance]: State of the random state |
| | at the end of minimization. |
| | |
| | For more details related to the OptimizeResult object, refer |
| | http://docs.scipy.org/doc/scipy/reference/generated/scipy.optimize.OptimizeResult.html |
| | """ |
| | specs = {"args": locals(), "function": "base_minimize"} |
| |
|
| | acq_optimizer_kwargs = { |
| | "n_points": n_points, |
| | "n_restarts_optimizer": n_restarts_optimizer, |
| | "n_jobs": n_jobs, |
| | } |
| | acq_func_kwargs = {"xi": xi, "kappa": kappa} |
| |
|
| | |
| | |
| |
|
| | |
| | if x0 is None: |
| | x0 = [] |
| | elif not isinstance(x0[0], (list, tuple)): |
| | x0 = [x0] |
| | if not isinstance(x0, list): |
| | raise ValueError("`x0` should be a list, but got %s" % type(x0)) |
| |
|
| | |
| | if n_random_starts is not None: |
| | warnings.warn( |
| | ( |
| | "n_random_starts will be removed in favour of " |
| | "n_initial_points. It overwrites n_initial_points." |
| | ), |
| | DeprecationWarning, |
| | ) |
| | n_initial_points = n_random_starts |
| |
|
| | if n_initial_points <= 0 and not x0: |
| | raise ValueError("Either set `n_initial_points` > 0," " or provide `x0`") |
| | |
| | if isinstance(y0, Iterable): |
| | y0 = list(y0) |
| | elif isinstance(y0, numbers.Number): |
| | y0 = [y0] |
| | required_calls = n_initial_points + (len(x0) if not y0 else 0) |
| | if n_calls < required_calls: |
| | raise ValueError("Expected `n_calls` >= %d, got %d" % (required_calls, n_calls)) |
| | |
| | n_random = n_initial_points |
| | n_initial_points = n_initial_points + len(x0) |
| |
|
| | |
| |
|
| | |
| | optimizer = Optimizer( |
| | dimensions, |
| | base_estimator, |
| | n_initial_points=n_initial_points, |
| | initial_point_generator=initial_point_generator, |
| | n_jobs=n_jobs, |
| | acq_func=acq_func, |
| | acq_optimizer=acq_optimizer, |
| | random_state=random_state, |
| | model_queue_size=model_queue_size, |
| | space_constraint=space_constraint, |
| | acq_optimizer_kwargs=acq_optimizer_kwargs, |
| | acq_func_kwargs=acq_func_kwargs, |
| | ) |
| | |
| | assert all(isinstance(p, Iterable) for p in x0) |
| | if not all(len(p) == optimizer.space.n_dims for p in x0): |
| | raise RuntimeError( |
| | "Optimization space (%s) and initial points in x0 " |
| | "use inconsistent dimensions." % optimizer.space |
| | ) |
| | |
| | callbacks = check_callback(callback) |
| | if verbose: |
| | callbacks.append( |
| | VerboseCallback( |
| | n_init=len(x0) if not y0 else 0, |
| | n_random=n_random, |
| | n_total=n_calls, |
| | ) |
| | ) |
| |
|
| | |
| |
|
| | |
| | result = None |
| | |
| | if x0 and y0 is None: |
| | y0 = list(map(func, x0)) |
| | n_calls -= len(y0) |
| | |
| | if x0: |
| | if not (isinstance(y0, Iterable) or isinstance(y0, numbers.Number)): |
| | raise ValueError( |
| | "`y0` should be an iterable or a scalar, got %s" % type(y0) |
| | ) |
| | if len(x0) != len(y0): |
| | raise ValueError("`x0` and `y0` should have the same length") |
| | result = optimizer.tell(x0, y0) |
| | result.specs = specs |
| | if eval_callbacks(callbacks, result): |
| | return result |
| |
|
| | |
| | for _ in range(n_calls): |
| | next_x = optimizer.ask() |
| | next_y = func(next_x) |
| | result = optimizer.tell(next_x, next_y) |
| | result.specs = specs |
| | if eval_callbacks(callbacks, result): |
| | break |
| |
|
| | return result |
| |
|