|
|
from collections import OrderedDict |
|
|
from copy import deepcopy |
|
|
from functools import wraps |
|
|
|
|
|
import numpy as np |
|
|
from joblib import dump as dump_ |
|
|
from joblib import load as load_ |
|
|
from scipy.optimize import OptimizeResult |
|
|
from scipy.optimize import minimize as sp_minimize |
|
|
from sklearn.base import is_regressor |
|
|
from sklearn.ensemble import GradientBoostingRegressor |
|
|
|
|
|
from .learning import ( |
|
|
ExtraTreesRegressor, |
|
|
GaussianProcessRegressor, |
|
|
GradientBoostingQuantileRegressor, |
|
|
RandomForestRegressor, |
|
|
) |
|
|
from .learning.gaussian_process.kernels import ConstantKernel, HammingKernel, Matern |
|
|
from .sampler import Grid, Halton, Hammersly, InitialPointGenerator, Lhs, Sobol |
|
|
from .space import Categorical, Dimension, Space |
|
|
|
|
|
__all__ = ( |
|
|
"load", |
|
|
"dump", |
|
|
) |
|
|
|
|
|
|
|
|
def create_result(Xi, yi, space=None, rng=None, specs=None, models=None): |
|
|
"""Initialize an `OptimizeResult` object. |
|
|
|
|
|
Parameters |
|
|
---------- |
|
|
Xi : list of lists, shape (n_iters, n_features) |
|
|
Location of the minimum at every iteration. |
|
|
|
|
|
yi : array-like, shape (n_iters,) |
|
|
Minimum value obtained at every iteration. |
|
|
|
|
|
space : Space instance, optional |
|
|
Search space. |
|
|
|
|
|
rng : RandomState instance, optional |
|
|
State of the random state. |
|
|
|
|
|
specs : dict, optional |
|
|
Call specifications. |
|
|
|
|
|
models : list, optional |
|
|
List of fit surrogate models. |
|
|
|
|
|
Returns |
|
|
------- |
|
|
res : `OptimizeResult`, scipy object |
|
|
OptimizeResult instance with the required information. |
|
|
""" |
|
|
res = OptimizeResult() |
|
|
yi = np.asarray(yi) |
|
|
if np.ndim(yi) == 2: |
|
|
res.log_time = np.ravel(yi[:, 1]) |
|
|
yi = np.ravel(yi[:, 0]) |
|
|
best = np.argmin(yi) |
|
|
res.x = Xi[best] |
|
|
res.fun = yi[best] |
|
|
res.func_vals = yi |
|
|
res.x_iters = Xi |
|
|
res.models = models |
|
|
res.space = space |
|
|
res.random_state = rng |
|
|
res.specs = specs |
|
|
return res |
|
|
|
|
|
|
|
|
def eval_callbacks(callbacks, result): |
|
|
"""Evaluate list of callbacks on result. |
|
|
|
|
|
The return values of the `callbacks` are ORed together to give the |
|
|
overall decision on whether or not the optimization procedure should |
|
|
continue. |
|
|
|
|
|
Parameters |
|
|
---------- |
|
|
callbacks : list of callables |
|
|
Callbacks to evaluate. |
|
|
|
|
|
result : `OptimizeResult`, scipy object |
|
|
Optimization result object to be stored. |
|
|
|
|
|
Returns |
|
|
------- |
|
|
decision : bool |
|
|
Decision of the callbacks whether or not to keep optimizing |
|
|
""" |
|
|
stop = False |
|
|
if callbacks: |
|
|
for c in callbacks: |
|
|
decision = c(result) |
|
|
if decision is not None: |
|
|
stop = stop or decision |
|
|
|
|
|
return stop |
|
|
|
|
|
|
|
|
def dump(res, filename, store_objective=True, **kwargs): |
|
|
"""Store an skopt optimization result into a file. |
|
|
|
|
|
Parameters |
|
|
---------- |
|
|
res : `OptimizeResult`, scipy object |
|
|
Optimization result object to be stored. |
|
|
|
|
|
filename : string or `pathlib.Path` |
|
|
The path of the file in which it is to be stored. The compression |
|
|
method corresponding to one of the supported filename extensions ('.z', |
|
|
'.gz', '.bz2', '.xz' or '.lzma') will be used automatically. |
|
|
|
|
|
store_objective : boolean, default=True |
|
|
Whether the objective function should be stored. Set `store_objective` |
|
|
to `False` if your objective function (`.specs['args']['func']`) is |
|
|
unserializable (i.e. if an exception is raised when trying to serialize |
|
|
the optimization result). |
|
|
|
|
|
Notice that if `store_objective` is set to `False`, a deep copy of the |
|
|
optimization result is created, potentially leading to performance |
|
|
problems if `res` is very large. If the objective function is not |
|
|
critical, one can delete it before calling `skopt.dump()` and thus |
|
|
avoid deep copying of `res`. |
|
|
|
|
|
**kwargs : other keyword arguments |
|
|
All other keyword arguments will be passed to `joblib.dump`. |
|
|
""" |
|
|
if store_objective: |
|
|
dump_(res, filename, **kwargs) |
|
|
|
|
|
elif 'func' in res.specs['args']: |
|
|
|
|
|
|
|
|
|
|
|
res_without_func = deepcopy(res) |
|
|
del res_without_func.specs['args']['func'] |
|
|
dump_(res_without_func, filename, **kwargs) |
|
|
|
|
|
else: |
|
|
|
|
|
|
|
|
dump_(res, filename, **kwargs) |
|
|
|
|
|
|
|
|
def load(filename, **kwargs): |
|
|
"""Reconstruct a skopt optimization result from a file persisted with skopt.dump. |
|
|
|
|
|
.. note:: |
|
|
Notice that the loaded optimization result can be missing |
|
|
the objective function (`.specs['args']['func']`) if `skopt.dump` |
|
|
was called with `store_objective=False`. |
|
|
|
|
|
Parameters |
|
|
---------- |
|
|
filename : string or `pathlib.Path` |
|
|
The path of the file from which to load the optimization result. |
|
|
|
|
|
**kwargs : other keyword arguments |
|
|
All other keyword arguments will be passed to `joblib.load`. |
|
|
|
|
|
Returns |
|
|
------- |
|
|
res : `OptimizeResult`, scipy object |
|
|
Reconstructed OptimizeResult instance. |
|
|
""" |
|
|
return load_(filename, **kwargs) |
|
|
|
|
|
|
|
|
def is_listlike(x): |
|
|
return isinstance(x, (list, tuple)) |
|
|
|
|
|
|
|
|
def is_2Dlistlike(x): |
|
|
return np.all([is_listlike(xi) for xi in x]) |
|
|
|
|
|
|
|
|
def check_x_in_space(x, space): |
|
|
if is_2Dlistlike(x): |
|
|
if not np.all([p in space for p in x]): |
|
|
raise ValueError("Not all points are within the bounds of" " the space.") |
|
|
if any([len(p) != len(space.dimensions) for p in x]): |
|
|
raise ValueError("Not all points have the same dimensions as" " the space.") |
|
|
elif is_listlike(x): |
|
|
if x not in space: |
|
|
raise ValueError( |
|
|
"Point (%s) is not within the bounds of" |
|
|
" the space (%s)." % (x, space.bounds) |
|
|
) |
|
|
if len(x) != len(space.dimensions): |
|
|
raise ValueError( |
|
|
"Dimensions of point (%s) and space (%s) do not match" |
|
|
% (x, space.bounds) |
|
|
) |
|
|
|
|
|
|
|
|
def expected_minimum(res, n_random_starts=20, random_state=None): |
|
|
"""Compute the minimum over the predictions of the last surrogate model. Uses |
|
|
`expected_minimum_random_sampling` with `n_random_starts` = 100000, when the space |
|
|
contains any categorical values. |
|
|
|
|
|
.. note:: |
|
|
The returned minimum may not necessarily be an accurate |
|
|
prediction of the minimum of the true objective function. |
|
|
|
|
|
Parameters |
|
|
---------- |
|
|
res : `OptimizeResult`, scipy object |
|
|
The optimization result returned by a `skopt` minimizer. |
|
|
|
|
|
n_random_starts : int, default=20 |
|
|
The number of random starts for the minimization of the surrogate |
|
|
model. |
|
|
|
|
|
random_state : int, RandomState instance, or None (default) |
|
|
Set random state to something other than None for reproducible |
|
|
results. |
|
|
|
|
|
Returns |
|
|
------- |
|
|
x : list |
|
|
location of the minimum. |
|
|
fun : float |
|
|
the surrogate function value at the minimum. |
|
|
""" |
|
|
if res.space.is_partly_categorical: |
|
|
return expected_minimum_random_sampling( |
|
|
res, n_random_starts=100000, random_state=random_state |
|
|
) |
|
|
|
|
|
def func(x): |
|
|
reg = res.models[-1] |
|
|
x = res.space.transform(x.reshape(1, -1)) |
|
|
return reg.predict(x.reshape(1, -1))[0] |
|
|
|
|
|
xs = [res.x] |
|
|
if n_random_starts > 0: |
|
|
xs.extend(res.space.rvs(n_random_starts, random_state=random_state)) |
|
|
|
|
|
best_x = None |
|
|
best_fun = np.inf |
|
|
|
|
|
for x0 in xs: |
|
|
r = sp_minimize(func, x0=x0, bounds=res.space.bounds) |
|
|
|
|
|
if r.fun < best_fun: |
|
|
best_x = r.x |
|
|
best_fun = r.fun |
|
|
|
|
|
return [v for v in best_x], best_fun |
|
|
|
|
|
|
|
|
def expected_minimum_random_sampling(res, n_random_starts=100000, random_state=None): |
|
|
"""Minimum search by doing naive random sampling, Returns the parameters that gave |
|
|
the minimum function value. Can be used when the space contains any categorical |
|
|
values. |
|
|
|
|
|
.. note:: |
|
|
The returned minimum may not necessarily be an accurate |
|
|
prediction of the minimum of the true objective function. |
|
|
|
|
|
Parameters |
|
|
---------- |
|
|
res : `OptimizeResult`, scipy object |
|
|
The optimization result returned by a `skopt` minimizer. |
|
|
|
|
|
n_random_starts : int, default=100000 |
|
|
The number of random starts for the minimization of the surrogate |
|
|
model. |
|
|
|
|
|
random_state : int, RandomState instance, or None (default) |
|
|
Set random state to something other than None for reproducible |
|
|
results. |
|
|
|
|
|
Returns |
|
|
------- |
|
|
x : list |
|
|
location of the minimum. |
|
|
fun : float |
|
|
the surrogate function value at the minimum. |
|
|
""" |
|
|
|
|
|
|
|
|
random_samples = res.space.rvs(n_random_starts, random_state=random_state) |
|
|
|
|
|
|
|
|
model = res.models[-1] |
|
|
y_random = model.predict(res.space.transform(random_samples)) |
|
|
index_best_objective = np.argmin(y_random) |
|
|
min_x = random_samples[index_best_objective] |
|
|
|
|
|
return min_x, y_random[index_best_objective] |
|
|
|
|
|
|
|
|
def has_gradients(estimator): |
|
|
"""Check if an estimator's ``predict`` method provides gradients. |
|
|
|
|
|
Parameters |
|
|
---------- |
|
|
estimator : |
|
|
sklearn BaseEstimator instance. |
|
|
""" |
|
|
tree_estimators = ( |
|
|
ExtraTreesRegressor, |
|
|
RandomForestRegressor, |
|
|
GradientBoostingQuantileRegressor, |
|
|
) |
|
|
|
|
|
|
|
|
if estimator is None: |
|
|
return False |
|
|
|
|
|
if isinstance(estimator, tree_estimators): |
|
|
return False |
|
|
|
|
|
categorical_gp = False |
|
|
if hasattr(estimator, "kernel"): |
|
|
params = estimator.get_params() |
|
|
categorical_gp = isinstance(estimator.kernel, HammingKernel) or any( |
|
|
[isinstance(params[p], HammingKernel) for p in params] |
|
|
) |
|
|
|
|
|
return not categorical_gp |
|
|
|
|
|
|
|
|
def cook_estimator(base_estimator, space=None, **kwargs): |
|
|
"""Cook a default estimator. |
|
|
|
|
|
For the special base_estimator called "DUMMY" the return value is None. |
|
|
This corresponds to sampling points at random, hence there is no need |
|
|
for an estimator. |
|
|
|
|
|
Parameters |
|
|
---------- |
|
|
base_estimator : "GP", "RF", "ET", "GBRT", "DUMMY" or sklearn regressor |
|
|
Should inherit from `sklearn.base.RegressorMixin`. |
|
|
In addition the `predict` method should have an optional `return_std` |
|
|
argument, which returns `std(Y | x)`` along with `E[Y | x]`. |
|
|
If base_estimator is one of ["GP", "RF", "ET", "GBRT", "DUMMY"], a |
|
|
surrogate model corresponding to the relevant `X_minimize` function |
|
|
is created. |
|
|
|
|
|
space : Space instance |
|
|
Has to be provided if the base_estimator is a gaussian process. |
|
|
Ignored otherwise. |
|
|
|
|
|
kwargs : dict |
|
|
Extra parameters provided to the base_estimator at init time. |
|
|
""" |
|
|
if isinstance(base_estimator, str): |
|
|
base_estimator = base_estimator.upper() |
|
|
if base_estimator not in ["GP", "ET", "RF", "GBRT", "DUMMY"]: |
|
|
raise ValueError( |
|
|
"Valid strings for the base_estimator parameter " |
|
|
" are: 'RF', 'ET', 'GP', 'GBRT' or 'DUMMY' not " |
|
|
"%s." % base_estimator |
|
|
) |
|
|
elif not is_regressor(base_estimator): |
|
|
raise ValueError("base_estimator has to be a regressor.") |
|
|
|
|
|
if base_estimator == "GP": |
|
|
if space is not None: |
|
|
space = Space(space) |
|
|
space = Space(normalize_dimensions(space.dimensions)) |
|
|
n_dims = space.transformed_n_dims |
|
|
is_cat = space.is_categorical |
|
|
else: |
|
|
raise ValueError("Expected a Space instance, not None.") |
|
|
|
|
|
cov_amplitude = ConstantKernel(1.0, (0.01, 1000.0)) |
|
|
|
|
|
if is_cat: |
|
|
other_kernel = HammingKernel(length_scale=np.ones(n_dims)) |
|
|
else: |
|
|
other_kernel = Matern( |
|
|
length_scale=np.ones(n_dims), |
|
|
length_scale_bounds=[(0.01, 100)] * n_dims, |
|
|
nu=2.5, |
|
|
) |
|
|
|
|
|
base_estimator = GaussianProcessRegressor( |
|
|
kernel=cov_amplitude * other_kernel, |
|
|
normalize_y=True, |
|
|
noise="gaussian", |
|
|
n_restarts_optimizer=2, |
|
|
) |
|
|
elif base_estimator == "RF": |
|
|
base_estimator = RandomForestRegressor(n_estimators=100, min_samples_leaf=3) |
|
|
elif base_estimator == "ET": |
|
|
base_estimator = ExtraTreesRegressor(n_estimators=100, min_samples_leaf=3) |
|
|
elif base_estimator == "GBRT": |
|
|
gbrt = GradientBoostingRegressor(n_estimators=30, loss="quantile") |
|
|
base_estimator = GradientBoostingQuantileRegressor(base_estimator=gbrt) |
|
|
|
|
|
elif base_estimator == "DUMMY": |
|
|
return None |
|
|
|
|
|
if ('n_jobs' in kwargs.keys()) and not hasattr(base_estimator, 'n_jobs'): |
|
|
del kwargs['n_jobs'] |
|
|
|
|
|
base_estimator.set_params(**kwargs) |
|
|
return base_estimator |
|
|
|
|
|
|
|
|
def cook_initial_point_generator(generator, **kwargs): |
|
|
"""Cook a default initial point generator. |
|
|
|
|
|
For the special generator called "random" the return value is None. |
|
|
|
|
|
Parameters |
|
|
---------- |
|
|
generator : "lhs", "sobol", "halton", "hammersly", "grid", "random" \ |
|
|
or InitialPointGenerator instance" |
|
|
Should inherit from `skopt.sampler.InitialPointGenerator`. |
|
|
|
|
|
kwargs : dict |
|
|
Extra parameters provided to the generator at init time. |
|
|
""" |
|
|
if generator is None: |
|
|
generator = "random" |
|
|
elif isinstance(generator, str): |
|
|
generator = generator.lower() |
|
|
if generator not in ["sobol", "halton", "hammersly", "lhs", "random", "grid"]: |
|
|
raise ValueError( |
|
|
"Valid strings for the generator parameter " |
|
|
" are: 'sobol', 'lhs', 'halton', 'hammersly'," |
|
|
"'random', or 'grid' not " |
|
|
"%s." % generator |
|
|
) |
|
|
elif not isinstance(generator, InitialPointGenerator): |
|
|
raise ValueError( |
|
|
"generator has to be an InitialPointGenerator." |
|
|
"Got %s" % (str(type(generator))) |
|
|
) |
|
|
|
|
|
if isinstance(generator, str): |
|
|
if generator == "sobol": |
|
|
generator = Sobol() |
|
|
elif generator == "halton": |
|
|
generator = Halton() |
|
|
elif generator == "hammersly": |
|
|
generator = Hammersly() |
|
|
elif generator == "lhs": |
|
|
generator = Lhs() |
|
|
elif generator == "grid": |
|
|
generator = Grid() |
|
|
elif generator == "random": |
|
|
return None |
|
|
generator.set_params(**kwargs) |
|
|
return generator |
|
|
|
|
|
|
|
|
def dimensions_aslist(search_space): |
|
|
"""Convert a dict representation of a search space into a list of dimensions, |
|
|
ordered by sorted(search_space.keys()). |
|
|
|
|
|
Parameters |
|
|
---------- |
|
|
search_space : dict |
|
|
Represents search space. The keys are dimension names (strings) |
|
|
and values are instances of classes that inherit from the class |
|
|
:class:`skopt.space.Dimension` (Real, Integer or Categorical) |
|
|
|
|
|
Returns |
|
|
------- |
|
|
params_space_list: list |
|
|
list of skopt.space.Dimension instances. |
|
|
|
|
|
Examples |
|
|
-------- |
|
|
>>> from skopt.space.space import Real, Integer |
|
|
>>> from skopt.utils import dimensions_aslist |
|
|
>>> search_space = {'name1': Real(0,1), |
|
|
... 'name2': Integer(2,4), 'name3': Real(-1,1)} |
|
|
>>> dimensions_aslist(search_space)[0] |
|
|
Real(low=0, high=1, prior='uniform', transform='identity') |
|
|
>>> dimensions_aslist(search_space)[1] |
|
|
Integer(low=2, high=4, prior='uniform', transform='identity') |
|
|
>>> dimensions_aslist(search_space)[2] |
|
|
Real(low=-1, high=1, prior='uniform', transform='identity') |
|
|
""" |
|
|
params_space_list = [search_space[k] for k in sorted(search_space.keys())] |
|
|
return params_space_list |
|
|
|
|
|
|
|
|
def point_asdict(search_space, point_as_list): |
|
|
"""Convert the list representation of a point from a search space to the dictionary |
|
|
representation, where keys are dimension names and values are corresponding to the |
|
|
values of dimensions in the list. |
|
|
|
|
|
.. seealso:: :class:`skopt.utils.point_aslist` |
|
|
|
|
|
Parameters |
|
|
---------- |
|
|
search_space : dict |
|
|
Represents search space. The keys are dimension names (strings) |
|
|
and values are instances of classes that inherit from the class |
|
|
:class:`skopt.space.Dimension` (Real, Integer or Categorical) |
|
|
|
|
|
point_as_list : list |
|
|
list with parameter values.The order of parameters in the list |
|
|
is given by sorted(params_space.keys()). |
|
|
|
|
|
Returns |
|
|
------- |
|
|
params_dict : OrderedDict |
|
|
dictionary with parameter names as keys to which |
|
|
corresponding parameter values are assigned. |
|
|
|
|
|
Examples |
|
|
-------- |
|
|
>>> from skopt.space.space import Real, Integer |
|
|
>>> from skopt.utils import point_asdict |
|
|
>>> search_space = {'name1': Real(0,1), |
|
|
... 'name2': Integer(2,4), 'name3': Real(-1,1)} |
|
|
>>> point_as_list = [0.66, 3, -0.15] |
|
|
>>> d = point_asdict(search_space, point_as_list) |
|
|
>>> d.keys(), d.values() |
|
|
(odict_keys(['name1', 'name2', 'name3']), odict_values([0.66, 3, -0.15])) |
|
|
""" |
|
|
params_dict = OrderedDict() |
|
|
for k, v in zip(sorted(search_space.keys()), point_as_list): |
|
|
params_dict[k] = v |
|
|
return params_dict |
|
|
|
|
|
|
|
|
def point_aslist(search_space, point_as_dict): |
|
|
"""Convert a dictionary representation of a point from a search space to the list |
|
|
representation. The list of values is created from the values of the dictionary, |
|
|
sorted by the names of dimensions used as keys. |
|
|
|
|
|
.. seealso:: :class:`skopt.utils.point_asdict` |
|
|
|
|
|
Parameters |
|
|
---------- |
|
|
search_space : dict |
|
|
Represents search space. The keys are dimension names (strings) |
|
|
and values are instances of classes that inherit from the class |
|
|
:class:`skopt.space.Dimension` (Real, Integer or Categorical) |
|
|
|
|
|
point_as_dict : dict |
|
|
dict with parameter names as keys to which corresponding |
|
|
parameter values are assigned. |
|
|
|
|
|
Returns |
|
|
------- |
|
|
point_as_list : list |
|
|
list with point values.The order of |
|
|
parameters in the list is given by sorted(params_space.keys()). |
|
|
|
|
|
Examples |
|
|
-------- |
|
|
>>> from skopt.space.space import Real, Integer |
|
|
>>> from skopt.utils import point_aslist |
|
|
>>> search_space = {'name1': Real(0,1), |
|
|
... 'name2': Integer(2,4), 'name3': Real(-1,1)} |
|
|
>>> point_as_dict = {'name1': 0.66, 'name2': 3, 'name3': -0.15} |
|
|
>>> point_aslist(search_space, point_as_dict) |
|
|
[0.66, 3, -0.15] |
|
|
""" |
|
|
point_as_list = [point_as_dict[k] for k in sorted(search_space.keys())] |
|
|
return point_as_list |
|
|
|
|
|
|
|
|
def normalize_dimensions(dimensions): |
|
|
"""Create a ``Space`` where all dimensions are normalized to unit range. |
|
|
|
|
|
This is particularly useful for Gaussian process based regressors and is |
|
|
used internally by ``gp_minimize``. |
|
|
|
|
|
Parameters |
|
|
---------- |
|
|
dimensions : list, shape (n_dims,) |
|
|
List of search space dimensions. |
|
|
Each search dimension can be defined either as |
|
|
|
|
|
- a `(lower_bound, upper_bound)` tuple (for `Real` or `Integer` |
|
|
dimensions), |
|
|
- a `(lower_bound, upper_bound, "prior")` tuple (for `Real` |
|
|
dimensions), |
|
|
- as a list of categories (for `Categorical` dimensions), or |
|
|
- an instance of a `Dimension` object (`Real`, `Integer` or |
|
|
`Categorical`). |
|
|
|
|
|
NOTE: The upper and lower bounds are inclusive for `Integer` |
|
|
dimensions. |
|
|
""" |
|
|
space = Space(dimensions) |
|
|
transformed_dimensions = [] |
|
|
for dimension in space.dimensions: |
|
|
if isinstance(dimension, Categorical): |
|
|
dimension.set_transformer("onehot") |
|
|
elif isinstance(dimension, Dimension): |
|
|
|
|
|
|
|
|
dimension.set_transformer("normalize") |
|
|
else: |
|
|
raise RuntimeError("Unknown dimension type " "(%s)" % type(dimension)) |
|
|
transformed_dimensions.append(dimension) |
|
|
|
|
|
return Space(transformed_dimensions) |
|
|
|
|
|
|
|
|
def check_list_types(x, types): |
|
|
"""Check whether all elements of a list `x` are of the correct type(s) and raise a |
|
|
ValueError if they are not. |
|
|
|
|
|
Note that `types` can be either a single object-type or a tuple |
|
|
of object-types. |
|
|
|
|
|
Raises `ValueError`, If one or more element in the list `x` is |
|
|
not of the correct type(s). |
|
|
|
|
|
Parameters |
|
|
---------- |
|
|
x : list |
|
|
List of objects. |
|
|
|
|
|
types : object or list(object) |
|
|
Either a single object-type or a tuple of object-types. |
|
|
""" |
|
|
|
|
|
|
|
|
err = list(filter(lambda a: not isinstance(a, types), x)) |
|
|
|
|
|
|
|
|
if len(err) > 0: |
|
|
msg = "All elements in list must be instances of {}, but found: {}" |
|
|
msg = msg.format(types, err) |
|
|
raise ValueError(msg) |
|
|
|
|
|
|
|
|
def check_dimension_names(dimensions): |
|
|
"""Check whether all dimensions have names. Raises `ValueError`, if one or more |
|
|
dimensions are unnamed. |
|
|
|
|
|
Parameters |
|
|
---------- |
|
|
dimensions : list(Dimension) |
|
|
List of Dimension-objects. |
|
|
""" |
|
|
|
|
|
|
|
|
err_dims = list(filter(lambda dim: dim.name is None, dimensions)) |
|
|
|
|
|
|
|
|
if len(err_dims) > 0: |
|
|
msg = "All dimensions must have names, but found: {}" |
|
|
msg = msg.format(err_dims) |
|
|
raise ValueError(msg) |
|
|
|
|
|
|
|
|
def use_named_args(dimensions): |
|
|
"""Wrapper / decorator for an objective function that uses named arguments to make |
|
|
it compatible with optimizers that use a single list of parameters. |
|
|
|
|
|
Your objective function can be defined as being callable using named |
|
|
arguments: `func(foo=123, bar=3.0, baz='hello')` for a search-space |
|
|
with dimensions named `['foo', 'bar', 'baz']`. But the optimizer |
|
|
will only pass a single list `x` of unnamed arguments when calling |
|
|
the objective function: `func(x=[123, 3.0, 'hello'])`. This wrapper |
|
|
converts your objective function with named arguments into one that |
|
|
accepts a list as argument, while doing the conversion automatically. |
|
|
|
|
|
The advantage of this is that you don't have to unpack the list of |
|
|
arguments `x` yourself, which makes the code easier to read and |
|
|
also reduces the risk of bugs if you change the number of dimensions |
|
|
or their order in the search-space. |
|
|
|
|
|
Examples |
|
|
-------- |
|
|
>>> # Define the search-space dimensions. They must all have names! |
|
|
>>> from skopt.space import Real |
|
|
>>> from skopt import forest_minimize |
|
|
>>> from skopt.utils import use_named_args |
|
|
>>> dim1 = Real(name='foo', low=0.0, high=1.0) |
|
|
>>> dim2 = Real(name='bar', low=0.0, high=1.0) |
|
|
>>> dim3 = Real(name='baz', low=0.0, high=1.0) |
|
|
>>> |
|
|
>>> # Gather the search-space dimensions in a list. |
|
|
>>> dimensions = [dim1, dim2, dim3] |
|
|
>>> |
|
|
>>> # Define the objective function with named arguments |
|
|
>>> # and use this function-decorator to specify the |
|
|
>>> # search-space dimensions. |
|
|
>>> @use_named_args(dimensions=dimensions) |
|
|
... def my_objective_function(foo, bar, baz): |
|
|
... return foo ** 2 + bar ** 4 + baz ** 8 |
|
|
>>> |
|
|
>>> # Not the function is callable from the outside as |
|
|
>>> # `my_objective_function(x)` where `x` is a list of unnamed arguments, |
|
|
>>> # which then wraps your objective function that is callable as |
|
|
>>> # `my_objective_function(foo, bar, baz)`. |
|
|
>>> # The conversion from a list `x` to named parameters `foo`, |
|
|
>>> # `bar`, `baz` |
|
|
>>> # is done automatically. |
|
|
>>> |
|
|
>>> # Run the optimizer on the wrapped objective function which is called |
|
|
>>> # as `my_objective_function(x)` as expected by `forest_minimize()`. |
|
|
>>> result = forest_minimize(func=my_objective_function, |
|
|
... dimensions=dimensions, |
|
|
... n_calls=20, base_estimator="ET", |
|
|
... random_state=4) |
|
|
>>> |
|
|
>>> # Print the best-found results in same format as the expected result. |
|
|
>>> print("Best fitness: " + str(result.fun)) |
|
|
Best fitness: 0.1948080835239698 |
|
|
>>> print("Best parameters: {}".format(result.x)) |
|
|
Best parameters: [0.44134853091052617, 0.06570954323368307, 0.17586123323419825] |
|
|
|
|
|
Parameters |
|
|
---------- |
|
|
dimensions : list(Dimension) |
|
|
List of `Dimension`-objects for the search-space dimensions. |
|
|
|
|
|
Returns |
|
|
------- |
|
|
wrapped_func : callable |
|
|
Wrapped objective function. |
|
|
""" |
|
|
|
|
|
def decorator(func): |
|
|
"""This uses more advanced Python features to wrap `func` using a function- |
|
|
decorator, which are not explained so well in the official Python documentation. |
|
|
|
|
|
A good video tutorial explaining how this works is found here: |
|
|
https://www.youtube.com/watch?v=KlBPCzcQNU8 |
|
|
|
|
|
Parameters |
|
|
---------- |
|
|
func : callable |
|
|
Function to minimize. Should take *named arguments* |
|
|
and return the objective value. |
|
|
""" |
|
|
|
|
|
|
|
|
check_list_types(dimensions, Dimension) |
|
|
|
|
|
|
|
|
check_dimension_names(dimensions) |
|
|
|
|
|
@wraps(func) |
|
|
def wrapper(x): |
|
|
"""This is the code that will be executed every time the wrapped / decorated |
|
|
`func` is being called. It takes `x` as a single list of parameters and |
|
|
converts them to named arguments and calls `func` with them. |
|
|
|
|
|
Parameters |
|
|
---------- |
|
|
x : list |
|
|
A single list of parameters e.g. `[123, 3.0, 'linear']` |
|
|
which will be converted to named arguments and passed |
|
|
to `func`. |
|
|
|
|
|
Returns |
|
|
------- |
|
|
objective_value |
|
|
The objective value returned by `func`. |
|
|
""" |
|
|
|
|
|
|
|
|
|
|
|
if len(x) != len(dimensions): |
|
|
msg = ( |
|
|
"Mismatch in number of search-space dimensions. " |
|
|
"len(dimensions)=={} and len(x)=={}" |
|
|
) |
|
|
msg = msg.format(len(dimensions), len(x)) |
|
|
raise ValueError(msg) |
|
|
|
|
|
|
|
|
|
|
|
arg_dict = {dim.name: value for dim, value in zip(dimensions, x)} |
|
|
|
|
|
|
|
|
objective_value = func(**arg_dict) |
|
|
|
|
|
return objective_value |
|
|
|
|
|
return wrapper |
|
|
|
|
|
return decorator |
|
|
|