import numpy as np
from math import inf
from scipy.optimize import shgo
from scipy.optimize import direct
from scipy.optimize import differential_evolution
import matplotlib.pyplot as plt
from spotpython.fun.objectivefunctions import Analytical
from spotpython.spot import Spot
from spotpython.utils.init import (fun_control_init, design_control_init, surrogate_control_init, optimizer_control_init)
from spotpython.data.diabetes import Diabetes
from spotpython.hyperdict.light_hyper_dict import LightHyperDict
from spotpython.fun.hyperlight import HyperLight
from spotpython.utils.eda import print_exp_table, print_res_table
12 Introduction to Sequential Parameter Optimization
The following libraries are used in this document:
This presents an introduction to spotpythons
’s Spot
class. The official spotpython
documentation can be found here: https://sequential-parameter-optimization.github.io/spotpython/.
12.1 Surrogate Model Based Optimization
Surrogate model based optimization methods are common approaches in simulation and optimization. The sequential parameter optimization toolbox (SPOT) was developed because there is a great need for sound statistical analysis of simulation and optimization algorithms (Bartz-Beielstein, Lasarczyk, and Preuss 2005). SPOT includes methods for tuning based on classical regression and analysis of variance techniques. It presents tree-based models such as classification and regression trees and random forests as well as Bayesian optimization (Gaussian process models, also known as Kriging). Combinations of different meta-modeling approaches are possible. SPOT comes with a sophisticated surrogate model based optimization method, that can handle discrete and continuous inputs. Furthermore, any model implemented in scikit-learn
can be used out-of-the-box as a surrogate in spotpython
.
SPOT implements key techniques such as exploratory fitness landscape analysis and sensitivity analysis. It can be used to understand the performance of various algorithms, while simultaneously giving insights into their algorithmic behavior.
The spot
loop consists of the following steps:
- Init: Build initial design \(X\) using the
design
method, e.g., Latin Hypercube Sampling (LHS). - Evaluate initial design on real objective \(f\): \(y = f(X)\) using the
fun
method, e.g.,fun_sphere
. - Build surrogate: \(S = S(X,y)\), using the
fit
method of the surrogate model, e.g.,Kriging
. - Optimize on surrogate: \(X_0 = \text{optimize}(S)\), using the
optimizer
method, e.g.,differential_evolution
. - Evaluate on real objective: \(y_0 = f(X_0)\), using the
fun
method from above. - Impute (Infill) new points: \(X = X \cup X_0\), \(y = y \cup y_0\), using
Spot
’sinfill
method. - Goto 3.
12.2 Advantages of the spotpython
Approach
Neural networks and many ML algorithms are non-deterministic, so results are noisy (i.e., depend on the the initialization of the weights). Enhanced noise handling strategies, OCBA (description from HPT-book).
Optimal Computational Budget Allocation (OCBA) is a very efficient solution to solve the “general ranking and selection problem” if the objective function is noisy. It allocates function evaluations in an uneven manner to identify the best solutions and to reduce the total optimization costs. [Chen10a, Bart11b] Given a total number of optimization samples \(N\) to be allocated to \(k\) competing solutions whose performance is depicted by random variables with means \(\bar{y}_i\) (\(i=1, 2, \ldots, k\)), and finite variances \(\sigma_i^2\), respectively, as \(N \to \infty\), the can be asymptotically maximized when \[\begin{align} \frac{N_i}{N_j} & = \left( \frac{ \sigma_i / \delta_{b,i}}{\sigma_j/ \delta_{b,j}} \right)^2, i,j \in \{ 1, 2, \ldots, k\}, \text{ and } i \neq j \neq b,\\ N_b &= \sigma_b \sqrt{ \sum_{i=1, i\neq b}^k \frac{N_i^2}{\sigma_i^2} }, \end{align}\] where \(N_i\) is the number of replications allocated to solution \(i\), \(\delta_{b,i} = \bar{y}_b - \bar{y}_i\), and \(\bar{y}_b \leq \min_{i\neq b} \bar{y}_i\) Bartz-Beielstein and Friese (2011).
Surrogate-based optimization: Better than grid search and random search (Reference to HPT-book)
Visualization
Importance based on the Kriging model
Sensitivity analysis. Exploratory fitness landscape analysis. Provides XAI methods (feature importance, integrated gradients, etc.)
Uncertainty quantification
Flexible, modular meta-modeling handling.
spotpython
comes with a Kriging model, which can be replaced by any model implemented inscikit-learn
.Enhanced metric handling, especially for categorical hyperparameters (any
sklearn
metric can be used). Default is..Integration with
TensorBoard
: Visualization of the hyperparameter tuning process, of the training steps, the model graph. Parallel coordinates plot, scatter plot matrix, and more.Reproducibility. Results are stored as pickle files. The results can be loaded and visualized at any time and be transferred between different machines and operating systems.
Handles
scikit-learn
models andPyTorch
models out-of-the-box. The user has to add a simple wrapper for passing the hyperparameters to use aPyTorch
model inspotpython
.Compatible with
Lightning
.User can add own models as plain python code.
User can add own data sets in various formats.
Flexible data handling and data preprocessing.
Many examples online (hyperparameter-tuning-cookbook).
spotpython
uses a robust optimizer that can even deal with hyperparameter-settings that cause crashes of the algorithms to be tuned.even if the optimum is not found, HPT with
spotpython
prevents the user from choosing bad hyperparameters in a systematic way (design of experiments).
12.3 Disadvantages of the spotpython Approach
- Time consuming
- Surrogate can be misguiding
12.4 An Initial Example: Optimization of the One-dimensional Sphere Function
12.4.1 The Objective Function: Sphere
The spotpython
package provides several classes of objective functions. We will use the sphere
-function, which is an analytical objective function. Analytical functions can be described by a (closed) formula: \[
f(x) = x^2.
\]
= Analytical().fun_sphere
fun = np.linspace(-1, 1, 100).reshape(-1,1)
x = fun(x)
y
plt.figure()"k")
plt.plot(x,y,
plt.grid()"x")
plt.xlabel("f(x)")
plt.ylabel("Sphere Function")
plt.title( plt.show()

The optimal solution of the sphere function is \(x^* = 0\) with \(f(x^*) = 0\).
12.4.2 The Spot
Method as an Optimization Algorithm Using a Surrogate Model
pyspot
implements the Spot
method. The Spot
method uses a surrogate model to approximate the objective function and to find the optimal solution by performing an optimization on the surrogate model. In this example, Spot
uses ten initial design points, which are sampled from a Latin Hypercube Sampling (LHS) design. The Spot
method will then build a surrogate model based on these initial design points and the corresponding objective function values. The surrogate model is then used to find the optimal solution. As a default, Spot
uses a Kriging surrogate model, which is a Gaussian process model. As a default, ten initial design points are sampled from a Latin Hypercube Sampling (LHS) design. Also as a default, 15 function evaluations are performed, i.e., the final surrogate model is built based on 15 points.
The specification of the lower
and upper
bounds of the input space is mandatory via the fun_control
dictionary, is passed to the Spot
method. For convenience, spotpython
provides an initialization method fun_control_init()
to create the fun_control
dictionary. After the Spot
method is initialized, the run()
method is called to start the optimization process. The run()
method will perform the optimization on the surrogate model and return the optimal solution. Finally, the print_results()
method is called to print the results of the optimization process, i.e., the best objective function value (“min y”) and the corresponding input value (“x0”).
=fun_control_init(lower = np.array([-1]),
fun_control= np.array([1]))
upper = Spot(fun=fun, fun_control=fun_control)
S
S.run() S.print_results()
Anisotropic model: n_theta set to 1
Anisotropic model: n_theta set to 1
spotpython tuning: 1.0155104003206277e-09 [#######---] 73.33%
Anisotropic model: n_theta set to 1
spotpython tuning: 1.0155104003206277e-09 [########--] 80.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 1.0155104003206277e-09 [#########-] 86.67%
Anisotropic model: n_theta set to 1
spotpython tuning: 1.0155104003206277e-09 [#########-] 93.33%
Anisotropic model: n_theta set to 1
spotpython tuning: 6.271448121765632e-11 [##########] 100.00% Done...
Experiment saved to 000_res.pkl
min y: 6.271448121765632e-11
x0: 7.919247515872725e-06
[['x0', np.float64(7.919247515872725e-06)]]
spotpython
provides a method plot_progress()
to visualize the search progress. The parameter log_y
is used to plot the objective function values on a logarithmic scale. The black points represent the ten initial design points, whereas the remainin g five red points represent the points found by the surrogate model based optimization.
=True) S.plot_progress(log_y
S.surrogate_control
{'log_level': 50,
'method': 'regression',
'model_optimizer': <function scipy.optimize._differentialevolution.differential_evolution(func, bounds, args=(), strategy='best1bin', maxiter=1000, popsize=15, tol=0.01, mutation=(0.5, 1), recombination=0.7, seed=None, callback=None, disp=False, polish=True, init='latinhypercube', atol=0, updating='immediate', workers=1, constraints=(), x0=None, *, integrality=None, vectorized=False)>,
'model_fun_evals': 10000,
'min_theta': -3.0,
'max_theta': 2.0,
'isotropic': False,
'p_val': 2.0,
'n_p': 1,
'optim_p': False,
'min_Lambda': -9,
'max_Lambda': 0,
'seed': 124,
'theta_init_zero': False,
'var_type': ['num'],
'metric_factorial': 'canberra'}
12.5 A Second Example: Optimization of a User-Specified Multi-dimensional Function
Users can easily specify their own objective functions. The following example shows how to use the Spot
method to optimize a user-specified two-dimensional function. Here we will use the 2-dimensional Branin function, which is a well-known test function in optimization. The Branin function is defined as follows: \[
f(x) = a (x2 - bx1^2 + cx1 - r)^2 + s(1-t) \cos(x1) + s
\] where:
- \(a = 1\),
- \(b = 5.1/(4\pi^2)\),
- \(c = 5/\pi\),
- \(r = 6\),
- \(s = 10\), and
- \(t = 1/(8\pi)\).
The input space is defined as \(x_1 \in [-5, 10]\) and \(x_2 \in [0, 15]\). The user specified Branin function can be implemented as follows (note, that the function is vectorized, i.e., it can handle multiple input points at once and that the **kwargs
argument is mandatory for compatibility with spotpython
to pass additional parameters to the function, if needed):
def user_fun(X, **kwargs):
= X[:, 0]
x1 = X[:, 1]
x2 = 1
a = 5.1 / (4 * np.pi**2)
b = 5 / np.pi
c = 6
r = 10
s = 1 / (8 * np.pi)
t = a * (x2 - b * x1**2 + c * x1 - r)**2 + s * (1 - t) * np.cos(x1) + s
y return y
The Branin function has its global minimum at three points: \(f(x^*) = 0.397887\), at \(x^* = (-\pi, 12.275)\), \((\pi, 2.275)\) and \((9.42478, 2.475)\). It can be visualized as shown in Figure 12.2.
= np.linspace(-5, 10, 100)
x1 = np.linspace(0, 15, 100)
x2 = np.meshgrid(x1, x2)
X1, X2 = user_fun(np.array([X1.ravel(), X2.ravel()]).T)
Z = Z.reshape(X1.shape)
Z =(8, 6))
plt.figure(figsize=30, cmap='jet')
plt.contourf(X1, X2, Z, levels='f(x)')
plt.colorbar(label'x1')
plt.xlabel('x2')
plt.ylabel('Branin Function')
plt.title(-np.pi, np.pi, 9.42478], [12.275, 2.275, 2.475], color='white', label='Global Minima')
plt.scatter([
plt.legend() plt.show()

The default Spot
class assumes a noisy objective function, because many real-world applications have noisy objective functions. Here, we will set the noise
parameter to False
in the fun_control
dictionary, because the Branin function is a deterministic function. Accordingly, the interpolation
surrogate method is used, which is suitable for deterministic functions. Therefore, we need to specify the interpolation
method in the surrogate_control
dictionary. Furthermore, since the Branin is harder to optimize than the sphere function, we will increase the number of function evaluations to 25 and the number of initial design points to 15.
= fun_control_init(
fun_control = np.array( [-5, 0]),
lower = np.array([10, 15]),
upper = 25,
fun_evals = False,
noise =123
seed
)
=design_control_init(init_size=15)
design_control=surrogate_control_init(method="interpolation")
surrogate_control
= Spot(fun=user_fun,
S_2 =fun_control,
fun_control=design_control,
design_control=surrogate_control)
surrogate_control
S_2.run()
S_2.print_results()=True) S_2.plot_progress(log_y
Anisotropic model: n_theta set to 2
Anisotropic model: n_theta set to 2
spotpython tuning: 3.1165078710223124 [######----] 64.00%
Anisotropic model: n_theta set to 2
spotpython tuning: 2.3234855354181514 [#######---] 68.00%
Anisotropic model: n_theta set to 2
spotpython tuning: 0.763617828201026 [#######---] 72.00%
Anisotropic model: n_theta set to 2
spotpython tuning: 0.763617828201026 [########--] 76.00%
Anisotropic model: n_theta set to 2
spotpython tuning: 0.7135503016263893 [########--] 80.00%
Anisotropic model: n_theta set to 2
spotpython tuning: 0.5268292655785878 [########--] 84.00%
Anisotropic model: n_theta set to 2
spotpython tuning: 0.4030140325970599 [#########-] 88.00%
Anisotropic model: n_theta set to 2
spotpython tuning: 0.39849132235934626 [#########-] 92.00%
Anisotropic model: n_theta set to 2
spotpython tuning: 0.3983320681394673 [##########] 96.00%
Anisotropic model: n_theta set to 2
spotpython tuning: 0.3983320681394673 [##########] 100.00% Done...
Experiment saved to 000_res.pkl
min y: 0.3983320681394673
x0: -3.1438152654671296
x1: 12.300860225578253
The resulting surrogate used by Spot
can be visualized as shown in Figure 12.3.
S_2.plot_contour()

S_2.surrogate_control
{'log_level': 50,
'method': 'interpolation',
'model_optimizer': <function scipy.optimize._differentialevolution.differential_evolution(func, bounds, args=(), strategy='best1bin', maxiter=1000, popsize=15, tol=0.01, mutation=(0.5, 1), recombination=0.7, seed=None, callback=None, disp=False, polish=True, init='latinhypercube', atol=0, updating='immediate', workers=1, constraints=(), x0=None, *, integrality=None, vectorized=False)>,
'model_fun_evals': 10000,
'min_theta': -3.0,
'max_theta': 2.0,
'isotropic': False,
'p_val': 2.0,
'n_p': 1,
'optim_p': False,
'min_Lambda': -9,
'max_Lambda': 0,
'seed': 124,
'theta_init_zero': False,
'var_type': ['num', 'num'],
'metric_factorial': 'canberra'}
12.6 A Third Example: Hyperparameter Tuning of a Neural Network
spotpython
provides several PyTorch
neural networks, which can be used out-of-the-box for hyperparameter tuning. We will use the Diabetes
dataset, which is a regression dataset with 10 input features and one output feature.
Similar to the steps from above, we define the fun_control
dictionary. The hyperparameter tuning of neural networks requires a few additional parameters in the fun_control
dictionary, e.g., for selecting the neural network model, the hyperparameter dictionary, and the number of input and output features of the neural network. The fun_control
dictionary is initialized using the fun_control_init()
method. The following parameters are used:
- The
core_model_name
parameter specifies the neural network model to be used. In this case, we will use theNNLinearRegressor
, which is a simple linear regression model implemented inspotpython
. - The
hyperdict
parameter specifies the hyperparameter dictionary to be used. In this case, we will use theLightHyperDict
, which is a hyperparameter dictionary for light-weight models. _L_in
and_L_out
specify the number of input and output features of the neural network, respectively. In this case, we will use 10 input features and 1 output feature, which are required for theDiabetes
dataset.
Sometimes, very bad configurations appear in the initial design, leading to an unnecessarily long optimization process. An example is illustrated in Figure 12.4. To avoid this, we can set a divergence threshold in the fun_control
dictionary. The divergence_threshold
parameter is used to stop the optimization process if the objective function value exceeds a certain threshold. This is useful to avoid long optimization processes if the objective function is not well-behaved. We have set the divergence_threshold=5_000
.

= fun_control_init(
fun_control ="S_3",
PREFIX=1,
max_time= Diabetes(),
data_set ="light.regression.NNLinearRegressor",
core_model_name=LightHyperDict,
hyperdict=5_000,
divergence_threshold=10,
_L_in=1)
_L_out
= Spot(fun=HyperLight().fun,
S_3 =fun_control)
fun_control S_3.surrogate_control
module_name: light
submodule_name: regression
model_name: NNLinearRegressor
{'log_level': 50,
'method': 'regression',
'model_optimizer': <function scipy.optimize._differentialevolution.differential_evolution(func, bounds, args=(), strategy='best1bin', maxiter=1000, popsize=15, tol=0.01, mutation=(0.5, 1), recombination=0.7, seed=None, callback=None, disp=False, polish=True, init='latinhypercube', atol=0, updating='immediate', workers=1, constraints=(), x0=None, *, integrality=None, vectorized=False)>,
'model_fun_evals': 10000,
'min_theta': -3.0,
'max_theta': 2.0,
'isotropic': False,
'p_val': 2.0,
'n_p': 1,
'optim_p': False,
'min_Lambda': -9,
'max_Lambda': 0,
'seed': 124,
'theta_init_zero': False,
'var_type': ['int',
'int',
'int',
'factor',
'factor',
'float',
'float',
'int',
'factor',
'factor'],
'metric_factorial': 'canberra'}
S_3.run()
S_3.print_results()=True) S_3.plot_progress(log_y
train_model result: {'val_loss': 13860.1416015625, 'hp_metric': 13860.1416015625}
train_model result: {'val_loss': 5115.26318359375, 'hp_metric': 5115.26318359375}
train_model result: {'val_loss': 5342.6416015625, 'hp_metric': 5342.6416015625}
train_model result: {'val_loss': 14101.1875, 'hp_metric': 14101.1875}
train_model result: {'val_loss': nan, 'hp_metric': nan}
train_model result: {'val_loss': 6.099777150087332e+18, 'hp_metric': 6.099777150087332e+18}
train_model result: {'val_loss': nan, 'hp_metric': nan}
train_model result: {'val_loss': 22807.197265625, 'hp_metric': 22807.197265625}
train_model result: {'val_loss': 11929.5634765625, 'hp_metric': 11929.5634765625}
train_model(): trainer.fit failed with exception: SparseAdam does not support dense gradients, please consider Adam instead
train_model result: {'val_loss': 24054.533203125, 'hp_metric': 24054.533203125}
Anisotropic model: n_theta set to 10
train_model(): trainer.fit failed with exception: SparseAdam does not support dense gradients, please consider Adam instead
train_model result: {'val_loss': 24064.412109375, 'hp_metric': 24064.412109375}
Anisotropic model: n_theta set to 10
spotpython tuning: 5115.26318359375 [######----] 60.00%
train_model result: {'val_loss': 23293.333984375, 'hp_metric': 23293.333984375}
Anisotropic model: n_theta set to 10
spotpython tuning: 5115.26318359375 [#######---] 66.67%
train_model result: {'val_loss': 21756.787109375, 'hp_metric': 21756.787109375}
Anisotropic model: n_theta set to 10
spotpython tuning: 5115.26318359375 [#######---] 73.33%
train_model result: {'val_loss': 15348.955078125, 'hp_metric': 15348.955078125}
Anisotropic model: n_theta set to 10
spotpython tuning: 5115.26318359375 [########--] 80.00%
train_model result: {'val_loss': 24008.998046875, 'hp_metric': 24008.998046875}
Anisotropic model: n_theta set to 10
spotpython tuning: 5115.26318359375 [#########-] 86.67%
train_model result: {'val_loss': 3099.8681640625, 'hp_metric': 3099.8681640625}
Anisotropic model: n_theta set to 10
spotpython tuning: 3099.8681640625 [#########-] 93.33%
Experiment saved to S_3_res.pkl
min y: 3099.8681640625
l1: 3.0
epochs: 5.0
batch_size: 1.0
act_fn: 3.0
optimizer: 10.0
dropout_prob: 0.008892951693496673
lr_mult: 0.8538886074471566
patience: 4.0
batch_norm: 0.0
initialization: 2.0
Results can be printed using Spot
’s print_results()
method. The results include the best objective function value, the corresponding input values, and the number of function evaluations.
= S_3.print_results() results
min y: 3099.8681640625
l1: 3.0
epochs: 5.0
batch_size: 1.0
act_fn: 3.0
optimizer: 10.0
dropout_prob: 0.008892951693496673
lr_mult: 0.8538886074471566
patience: 4.0
batch_norm: 0.0
initialization: 2.0
S_3.surrogate_control
{'log_level': 50,
'method': 'regression',
'model_optimizer': <function scipy.optimize._differentialevolution.differential_evolution(func, bounds, args=(), strategy='best1bin', maxiter=1000, popsize=15, tol=0.01, mutation=(0.5, 1), recombination=0.7, seed=None, callback=None, disp=False, polish=True, init='latinhypercube', atol=0, updating='immediate', workers=1, constraints=(), x0=None, *, integrality=None, vectorized=False)>,
'model_fun_evals': 10000,
'min_theta': -3.0,
'max_theta': 2.0,
'isotropic': False,
'p_val': 2.0,
'n_p': 1,
'optim_p': False,
'min_Lambda': -9,
'max_Lambda': 0,
'seed': 124,
'theta_init_zero': False,
'var_type': ['int',
'int',
'int',
'factor',
'factor',
'float',
'float',
'int',
'factor',
'factor'],
'metric_factorial': 'canberra'}
A formatted table of the results can be printed using the print_res_table()
method:
print_res_table(S_3)
| name | type | default | lower | upper | tuned | transform | importance | stars |
|----------------|--------|-----------|---------|---------|----------------------|-----------------------|--------------|---------|
| l1 | int | 3 | 3.0 | 8.0 | 3.0 | transform_power_2_int | 0.00 | |
| epochs | int | 4 | 4.0 | 9.0 | 5.0 | transform_power_2_int | 100.00 | *** |
| batch_size | int | 4 | 1.0 | 4.0 | 1.0 | transform_power_2_int | 0.00 | |
| act_fn | factor | ReLU | 0.0 | 5.0 | LeakyReLU | None | 0.00 | |
| optimizer | factor | SGD | 0.0 | 11.0 | Rprop | None | 0.00 | |
| dropout_prob | float | 0.01 | 0.0 | 0.25 | 0.008892951693496673 | None | 0.00 | |
| lr_mult | float | 1.0 | 0.1 | 10.0 | 0.8538886074471566 | None | 0.00 | |
| patience | int | 2 | 2.0 | 6.0 | 4.0 | transform_power_2_int | 0.00 | |
| batch_norm | factor | 0 | 0.0 | 1.0 | 0 | None | 23.73 | * |
| initialization | factor | Default | 0.0 | 4.0 | kaiming_normal | None | 0.00 | |
The fitness landscape can be visualized using the plot_important_hyperparameter_contour()
method:
=3) S_3.plot_important_hyperparameter_contour(max_imp
l1: 0.001101318314260883
epochs: 100.00000000000001
batch_size: 0.001101318314260883
act_fn: 0.001101318314260883
optimizer: 0.001101318314260883
dropout_prob: 0.001101318314260883
lr_mult: 0.001101318314260883
patience: 0.001101318314260883
batch_norm: 23.731435390281487
initialization: 0.001101318314260883
12.7 Organization
Spot
organizes the surrogate based optimization process in four steps:
- Selection of the objective function:
fun
. - Selection of the initial design:
design
. - Selection of the optimization algorithm:
optimizer
. - Selection of the surrogate model:
surrogate
.
For each of these steps, the user can specify an object:
from spotpython.fun.objectivefunctions import Analytical
= Analytical().fun_sphere
fun from spotpython.design.spacefilling import SpaceFilling
= SpaceFilling(2)
design from scipy.optimize import differential_evolution
= differential_evolution
optimizer from spotpython.surrogate.kriging import Kriging
= Kriging() surrogate
For each of these steps, the user can specify a dictionary of control parameters.
fun_control
design_control
optimizer_control
surrogate_control
Each of these dictionaries has an initialization method, e.g., fun_control_init()
. The initialization methods set the default values for the control parameters.
- The specification of an lower bound in
fun_control
is mandatory.
12.8 Summary: Basic Attributes and Methods of the Spot
Object
The Spot
class organizes the surrogate based optimization process into the
fun
object with dictionryfun_control
,design
object with dictionarydesign_control
,optimizer
object with dictionaryoptimizer_control
, andsurrogate
object with dictionarysurrogate_control
.
Object | Description | Default |
---|---|---|
fun |
Can be one of the following: (1) any function from spotpython.fun.* , e.g., fun_rosen, (2) a user-specified function (vectorized and accepting the **kwargs argument as, e.g., in Section 12.5), or, (3) if PyTorch hyperparameter tuning with Lightning is desired, the HyperLight class from spotpython , see HyperLight. |
There is no default, a simple example is fun = Analytical().fun_sphere |
design |
Can be any design generating function from spotpython.design:* , e.g., Clustered, Factorial, Grid, Poor, Random, Sobol, or SpaceFilling. |
design = SpaceFilling(k=15) |
optimizer |
Can be any of dual_annealing , direct , shgo , basinhopping , or differential_evolution , see, e.g., differential_evolution. |
optimizer=differential_evolution |
surrogate |
Can be any sklearn regressor, e.g., GaussianProcessRegressor or, the Kriging class from spotpython . |
surrogate=Kriging() |
Dictionary | Code Reference | Example |
---|---|---|
fun_control |
fun_control_init | The lower and upper arguments are mandatory, e.g., fun_control=fun_control_init(lower=np.array([-1, -1]),upper=np.array([1, 1])) . |
design_control |
design_control_init() | design_control=design_control_init() |
optimizer_control |
optimizer_control_init | optimizer_control_init() |
surrogate_control |
surrogate_control_init | surrogate_control=surrogate_control_init() |
Based on the definition of the fun
, design
, optimizer
, and surrogate
objects, and their corresponding control parameter dictionaries, fun_control
, design_control
, optimizer_control
, and surrogate_control
, the Spot
object can be build as follows:
=fun_control_init(lower=np.array([-1, -1]),
fun_control=np.array([1, 1]))
upper=design_control_init()
design_control=optimizer_control_init()
optimizer_control=surrogate_control_init()
surrogate_control= Spot(fun=fun,
spot_tuner =fun_control,
fun_control=design_control,
design_control=optimizer_control,
optimizer_control=surrogate_control) surrogate_control
12.9 Run
spot_tuner.run()
Anisotropic model: n_theta set to 2
Anisotropic model: n_theta set to 2
spotpython tuning: 7.295281981238084e-06 [#######---] 73.33%
Anisotropic model: n_theta set to 2
spotpython tuning: 7.295281981238084e-06 [########--] 80.00%
Anisotropic model: n_theta set to 2
spotpython tuning: 7.295281981238084e-06 [#########-] 86.67%
Anisotropic model: n_theta set to 2
spotpython tuning: 7.295281981238084e-06 [#########-] 93.33%
Anisotropic model: n_theta set to 2
spotpython tuning: 7.295281981238084e-06 [##########] 100.00% Done...
Experiment saved to 000_res.pkl
<spotpython.spot.spot.Spot at 0x3a6eb1a00>
12.10 Print the Results
spot_tuner.print_results()
min y: 7.295281981238084e-06
x0: 0.0005342902641415708
x1: 0.0026476056909746993
[['x0', np.float64(0.0005342902641415708)],
['x1', np.float64(0.0026476056909746993)]]
12.11 Show the Progress
spot_tuner.plot_progress()
12.12 Visualize the Surrogate
- The plot method of the
kriging
surrogate is used. - Note: the plot uses the interval defined by the ranges of the natural variables.
spot_tuner.surrogate.plot()
12.13 Init: Build Initial Design
from spotpython.design.spacefilling import SpaceFilling
from spotpython.surrogate.kriging import Kriging
from spotpython.fun.objectivefunctions import Analytical
from spotpython.utils.init import fun_control_init
= SpaceFilling(2)
gen = np.random.RandomState(1)
rng = np.array([-5,-0])
lower = np.array([10,15])
upper = Analytical().fun_branin
fun
= fun_control_init(sigma=0)
fun_control
= gen.scipy_lhd(10, lower=lower, upper = upper)
X print(X)
= fun(X, fun_control=fun_control)
y print(y)
[[ 8.97647221 13.41926847]
[ 0.66946019 1.22344228]
[ 5.23614115 13.78185824]
[ 5.6149825 11.5851384 ]
[-1.72963184 1.66516096]
[-4.26945568 7.1325531 ]
[ 1.26363761 10.17935555]
[ 2.88779942 8.05508969]
[-3.39111089 4.15213772]
[ 7.30131231 5.22275244]]
[128.95676449 31.73474356 172.89678121 126.71295908 64.34349975
70.16178611 48.71407916 31.77322887 76.91788181 30.69410529]
12.14 Replicability
Seed
= SpaceFilling(2, seed=123)
gen = gen.scipy_lhd(3)
X0 = SpaceFilling(2, seed=345)
gen = gen.scipy_lhd(3)
X1 = gen.scipy_lhd(3)
X2 = SpaceFilling(2, seed=123)
gen = gen.scipy_lhd(3)
X3 X0, X1, X2, X3
(array([[0.77254938, 0.31539299],
[0.59321338, 0.93854273],
[0.27469803, 0.3959685 ]]),
array([[0.78373509, 0.86811887],
[0.06692621, 0.6058029 ],
[0.41374778, 0.00525456]]),
array([[0.121357 , 0.69043832],
[0.41906219, 0.32838498],
[0.86742658, 0.52910374]]),
array([[0.77254938, 0.31539299],
[0.59321338, 0.93854273],
[0.27469803, 0.3959685 ]]))
12.15 Surrogates
12.15.1 A Simple Predictor
The code below shows how to use a simple model for prediction. Assume that only two (very costly) measurements are available:
- f(0) = 0.5
- f(2) = 2.5
We are interested in the value at \(x_0 = 1\), i.e., \(f(x_0 = 1)\), but cannot run an additional, third experiment.
from sklearn import linear_model
= np.array([[0], [2]])
X = np.array([0.5, 2.5])
y = linear_model.LinearRegression()
S_lm = S_lm.fit(X, y)
S_lm = np.array([[1]])
X0 = S_lm.predict(X0)
y0 print(y0)
[1.5]
Central Idea: Evaluation of the surrogate model S_lm
is much cheaper (or / and much faster) than running the real-world experiment \(f\).
12.16 Tensorboard Setup
12.16.1 Tensorboard Configuration
The TENSORBOARD_CLEAN
argument can be set to True
in the fun_control
dictionary to archive the TensorBoard folder if it already exists. This is useful if you want to start a hyperparameter tuning process from scratch. If you want to continue a hyperparameter tuning process, set TENSORBOARD_CLEAN
to False
. Then the TensorBoard folder will not be archived and the old and new TensorBoard files will shown in the TensorBoard dashboard.
12.16.2 Starting TensorBoard
TensorBoard
can be started as a background process with the following command, where ./runs
is the default directory for the TensorBoard log files:
tensorboard --logdir="./runs"
The TensorBoard path can be printed with the following command (after a fun_control
object has been created):
from spotpython.utils.init import get_tensorboard_path
get_tensorboard_path(fun_control)
12.17 Demo/Test: Objective Function Fails
SPOT expects np.nan
values from failed objective function values. These are handled. Note: SPOT’s counter considers only successful executions of the objective function.
import numpy as np
from spotpython.fun.objectivefunctions import Analytical
from spotpython.spot import Spot
import numpy as np
from math import inf
# number of initial points:
= 20
ni # number of points
= 30
n
= Analytical().fun_random_error
fun =fun_control_init(
fun_control= np.array([-1]),
lower = np.array([1]),
upper= n,
fun_evals =False)
show_progress=design_control_init(init_size=ni)
design_control
= Spot(fun=fun,
spot_1 =fun_control,
fun_control=design_control)
design_control
# assert value error from the run method
try:
spot_1.run()except ValueError as e:
print(e)
Anisotropic model: n_theta set to 1
Anisotropic model: n_theta set to 1
Anisotropic model: n_theta set to 1
Anisotropic model: n_theta set to 1
Anisotropic model: n_theta set to 1
Anisotropic model: n_theta set to 1
Anisotropic model: n_theta set to 1
Anisotropic model: n_theta set to 1
Anisotropic model: n_theta set to 1
Anisotropic model: n_theta set to 1
Anisotropic model: n_theta set to 1
Anisotropic model: n_theta set to 1
Anisotropic model: n_theta set to 1
Experiment saved to 000_res.pkl
12.18 Handling Results: Printing, Saving, and Loading
The results can be printed with the following command:
=False) spot_tuner.print_results(print_screen
The tuned hyperparameters can be obtained as a dictionary with the following command:
from spotpython.hyperparameters.values import get_tuned_hyperparameters
get_tuned_hyperparameters(spot_tuner, fun_control)
The results can be saved and reloaded with the following commands:
from spotpython.utils.file import save_pickle, load_pickle
from spotpython.utils.init import get_experiment_name
= get_experiment_name("024")
experiment_name = False
SAVE_AND_LOAD if SAVE_AND_LOAD == True:
save_pickle(spot_tuner, experiment_name)= load_pickle(experiment_name) spot_tuner
12.19 spotpython
as a Hyperparameter Tuner
12.19.1 Modifying Hyperparameter Levels
spotpython
distinguishes between different types of hyperparameters. The following types are supported:
int
(integer)float
(floating point number)boolean
(boolean)factor
(categorical)
12.19.1.1 Integer Hyperparameters
Integer hyperparameters can be modified with the set_int_hyperparameter_values()
[SOURCE] function. The following code snippet shows how to modify the n_estimators
hyperparameter of a random forest model:
from spotriver.hyperdict.river_hyper_dict import RiverHyperDict
from spotpython.utils.init import fun_control_init
from spotpython.hyperparameters.values import set_int_hyperparameter_values
from spotpython.utils.eda import print_exp_table
= fun_control_init(
fun_control ="forest.AMFRegressor",
core_model_name=RiverHyperDict,
hyperdict
)print("Before modification:")
print_exp_table(fun_control)"n_estimators", 2, 5)
set_int_hyperparameter_values(fun_control, print("After modification:")
print_exp_table(fun_control)
Before modification:
| name | type | default | lower | upper | transform |
|-----------------|--------|-----------|---------|---------|-----------------------|
| n_estimators | int | 3 | 2 | 5 | transform_power_2_int |
| step | float | 1 | 0.1 | 10 | None |
| use_aggregation | factor | 1 | 0 | 1 | None |
Setting hyperparameter n_estimators to value [2, 5].
Variable type is int.
Core type is None.
Calling modify_hyper_parameter_bounds().
After modification:
| name | type | default | lower | upper | transform |
|-----------------|--------|-----------|---------|---------|-----------------------|
| n_estimators | int | 3 | 2 | 5 | transform_power_2_int |
| step | float | 1 | 0.1 | 10 | None |
| use_aggregation | factor | 1 | 0 | 1 | None |
12.19.1.2 Float Hyperparameters
Float hyperparameters can be modified with the set_float_hyperparameter_values()
[SOURCE] function. The following code snippet shows how to modify the step
hyperparameter of a hyperparameter of a Mondrian Regression Tree model:
from spotriver.hyperdict.river_hyper_dict import RiverHyperDict
from spotpython.utils.init import fun_control_init
from spotpython.hyperparameters.values import set_float_hyperparameter_values
from spotpython.utils.eda import print_exp_table
= fun_control_init(
fun_control ="forest.AMFRegressor",
core_model_name=RiverHyperDict,
hyperdict
)print("Before modification:")
print_exp_table(fun_control)"step", 0.2, 5)
set_float_hyperparameter_values(fun_control, print("After modification:")
print_exp_table(fun_control)
Before modification:
| name | type | default | lower | upper | transform |
|-----------------|--------|-----------|---------|---------|-----------------------|
| n_estimators | int | 3 | 2 | 5 | transform_power_2_int |
| step | float | 1 | 0.1 | 10 | None |
| use_aggregation | factor | 1 | 0 | 1 | None |
Setting hyperparameter step to value [0.2, 5].
Variable type is float.
Core type is None.
Calling modify_hyper_parameter_bounds().
After modification:
| name | type | default | lower | upper | transform |
|-----------------|--------|-----------|---------|---------|-----------------------|
| n_estimators | int | 3 | 2 | 5 | transform_power_2_int |
| step | float | 1 | 0.2 | 5 | None |
| use_aggregation | factor | 1 | 0 | 1 | None |
12.19.1.3 Boolean Hyperparameters
Boolean hyperparameters can be modified with the set_boolean_hyperparameter_values()
[SOURCE] function. The following code snippet shows how to modify the use_aggregation
hyperparameter of a Mondrian Regression Tree model:
from spotriver.hyperdict.river_hyper_dict import RiverHyperDict
from spotpython.utils.init import fun_control_init
from spotpython.hyperparameters.values import set_boolean_hyperparameter_values
from spotpython.utils.eda import print_exp_table
= fun_control_init(
fun_control ="forest.AMFRegressor",
core_model_name=RiverHyperDict,
hyperdict
)print("Before modification:")
print_exp_table(fun_control)"use_aggregation", 0, 0)
set_boolean_hyperparameter_values(fun_control, print("After modification:")
print_exp_table(fun_control)
Before modification:
| name | type | default | lower | upper | transform |
|-----------------|--------|-----------|---------|---------|-----------------------|
| n_estimators | int | 3 | 2 | 5 | transform_power_2_int |
| step | float | 1 | 0.1 | 10 | None |
| use_aggregation | factor | 1 | 0 | 1 | None |
Setting hyperparameter use_aggregation to value [0, 0].
Variable type is factor.
Core type is bool.
Calling modify_boolean_hyper_parameter_levels().
After modification:
| name | type | default | lower | upper | transform |
|-----------------|--------|-----------|---------|---------|-----------------------|
| n_estimators | int | 3 | 2 | 5 | transform_power_2_int |
| step | float | 1 | 0.1 | 10 | None |
| use_aggregation | factor | 1 | 0 | 0 | None |
12.19.1.4 Factor Hyperparameters
Factor hyperparameters can be modified with the set_factor_hyperparameter_values()
[SOURCE] function. The following code snippet shows how to modify the leaf_model
hyperparameter of a Hoeffding Tree Regressor model:
from spotriver.hyperdict.river_hyper_dict import RiverHyperDict
from spotpython.utils.init import fun_control_init
from spotpython.hyperparameters.values import set_factor_hyperparameter_values
from spotpython.utils.eda import print_exp_table
= fun_control_init(
fun_control ="tree.HoeffdingTreeRegressor",
core_model_name=RiverHyperDict,
hyperdict
)print("Before modification:")
print_exp_table(fun_control)"leaf_model", ['LinearRegression',
set_factor_hyperparameter_values(fun_control, 'Perceptron'])
print("After modification:")
Before modification:
| name | type | default | lower | upper | transform |
|------------------------|--------|------------------|---------|----------|------------------------|
| grace_period | int | 200 | 10 | 1000 | None |
| max_depth | int | 20 | 2 | 20 | transform_power_2_int |
| delta | float | 1e-07 | 1e-08 | 1e-06 | None |
| tau | float | 0.05 | 0.01 | 0.1 | None |
| leaf_prediction | factor | mean | 0 | 2 | None |
| leaf_model | factor | LinearRegression | 0 | 2 | None |
| model_selector_decay | float | 0.95 | 0.9 | 0.99 | None |
| splitter | factor | EBSTSplitter | 0 | 2 | None |
| min_samples_split | int | 5 | 2 | 10 | None |
| binary_split | factor | 0 | 0 | 1 | None |
| max_size | float | 500.0 | 100 | 1000 | None |
| memory_estimate_period | int | 6 | 3 | 8 | transform_power_10_int |
| stop_mem_management | factor | 0 | 0 | 1 | None |
| remove_poor_attrs | factor | 0 | 0 | 1 | None |
| merit_preprune | factor | 1 | 0 | 1 | None |
After modification:
12.20 Sampling in spotpython
spotpython
uses a class for generating space-filling designs using Latin Hypercube Sampling (LHS) and maximin distance criteria. It is based on scipy
’s LatinHypercube
class. The following example demonstrates how to generate a Latin Hypercube Sampling design using spotpython
. The result is shown in Figure 12.5. As can seen in the figure, a Latin hypercube sample generates \(n\) points in \([0,1)^{d}\). Each univariate marginal distribution is stratified, placing exactly one point in \([j/n, (j+1)/n)\) for \(j=0,1,...,n-1\).
import matplotlib.pyplot as plt
import numpy as np
from spotpython.design.spacefilling import SpaceFilling
= SpaceFilling(k=2, seed=123)
lhd = lhd.scipy_lhd(n=10, repeats=1, lower=np.array([0, 0]), upper=np.array([10, 10]))
X 0], X[:, 1])
plt.scatter(X[:, 'x1')
plt.xlabel('x2')
plt.ylabel( plt.grid()

12.21 Example: Spot
and the Sphere Function
Central Idea: Evaluation of the surrogate model S
is much cheaper (or / and much faster) than running the real-world experiment \(f\). We start with a small example.
import numpy as np
from math import inf
from spotpython.fun.objectivefunctions import Analytical
from spotpython.utils.init import fun_control_init, design_control_init
from spotpython.hyperparameters.values import set_control_key_value
from spotpython.spot import Spot
import matplotlib.pyplot as plt
12.21.1 The Objective Function: Sphere
The spotpython
package provides several classes of objective functions. We will use an analytical objective function, i.e., a function that can be described by a (closed) formula: \[
f(x) = x^2
\]
= Analytical().fun_sphere fun
We can apply the function fun
to input values and plot the result:
= np.linspace(-1,1,100).reshape(-1,1)
x = fun(x)
y
plt.figure()"k")
plt.plot(x, y, plt.show()
12.21.2 The Spot
Method as an Optimization Algorithm Using a Surrogate Model
We initialize the fun_control
dictionary. The fun_control
dictionary contains the parameters for the objective function. The fun_control
dictionary is passed to the Spot
method.
=fun_control_init(lower = np.array([-1]),
fun_control= np.array([1]))
upper = Spot(fun=fun,
spot_0 =fun_control)
fun_control spot_0.run()
Anisotropic model: n_theta set to 1
Anisotropic model: n_theta set to 1
spotpython tuning: 1.0155104003206277e-09 [#######---] 73.33%
Anisotropic model: n_theta set to 1
spotpython tuning: 1.0155104003206277e-09 [########--] 80.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 1.0155104003206277e-09 [#########-] 86.67%
Anisotropic model: n_theta set to 1
spotpython tuning: 1.0155104003206277e-09 [#########-] 93.33%
Anisotropic model: n_theta set to 1
spotpython tuning: 6.271448121765632e-11 [##########] 100.00% Done...
Experiment saved to 000_res.pkl
<spotpython.spot.spot.Spot at 0x1629987a0>
The method print_results()
prints the results, i.e., the best objective function value (“min y”) and the corresponding input value (“x0”).
spot_0.print_results()
min y: 6.271448121765632e-11
x0: 7.919247515872725e-06
[['x0', np.float64(7.919247515872725e-06)]]
To plot the search progress, the method plot_progress()
can be used. The parameter log_y
is used to plot the objective function values on a logarithmic scale.
=True) spot_0.plot_progress(log_y

Spot
method. The black elements (points and line) represent the initial design, before the surrogate is build. The red elements represent the search on the surrogate.
If the dimension of the input space is one, the method plot_model()
can be used to visualize the model and the underlying objective function values.
spot_0.plot_model()

12.22 Spot
Parameters: fun_evals
, init_size
and show_models
We will modify three parameters:
- The number of function evaluations (
fun_evals
) will be set to10
(instead of 15, which is the default value) in thefun_control
dictionary. - The parameter
show_models
, which visualizes the search process for each single iteration for 1-dim functions, in thefun_control
dictionary. - The size of the initial design (
init_size
) in thedesign_control
dictionary.
The full list of the Spot
parameters is shown in code reference on GitHub, see Spot.
=fun_control_init(lower = np.array([-1]),
fun_control= np.array([1]),
upper = 10,
fun_evals = True)
show_models = design_control_init(init_size=9)
design_control = Spot(fun=fun,
spot_1 =fun_control,
fun_control=design_control)
design_control spot_1.run()
Anisotropic model: n_theta set to 1
Anisotropic model: n_theta set to 1
spotpython tuning: 4.960036375406414e-09 [##########] 100.00% Done...
Experiment saved to 000_res.pkl
12.23 Print the Results
spot_1.print_results()
min y: 4.960036375406414e-09
x0: -7.042752569419441e-05
[['x0', np.float64(-7.042752569419441e-05)]]
12.24 Show the Progress
spot_1.plot_progress()
12.25 Visualizing the Optimization and Hyperparameter Tuning Process with TensorBoard
spotpython
supports the visualization of the hyperparameter tuning process with TensorBoard. The following example shows how to use TensorBoard with spotpython
.
First, we define an “PREFIX” to identify the hyperparameter tuning process. The PREFIX is used to create a directory for the TensorBoard files.
= fun_control_init(
fun_control = "01",
PREFIX = np.array([-1]),
lower = np.array([2]),
upper =100,
fun_evals=True,
TENSORBOARD_CLEAN=True)
tensorboard_log= design_control_init(init_size=5) design_control
Moving TENSORBOARD_PATH: runs/ to TENSORBOARD_PATH_OLD: runs_OLD/runs_2025_08_07_17_49_40_0
Created spot_tensorboard_path: runs/spot_logs/01_maans08_2025-08-07_17-49-40 for SummaryWriter()
Since the tensorboard_log
is True
, spotpython
will log the optimization process in the TensorBoard files. The argument TENSORBOARD_CLEAN=True
will move the TensorBoard files from the previous run to a backup folder, so that TensorBoard files from previous runs are not overwritten and a clean start in the runs
folder is guaranteed.
= Spot(fun=fun,
spot_tuner =fun_control,
fun_control=design_control)
design_control
spot_tuner.run() spot_tuner.print_results()
Anisotropic model: n_theta set to 1
Anisotropic model: n_theta set to 1
spotpython tuning: 6.428084284497196e-06 [#---------] 6.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 1.0020061415058982e-07 [#---------] 7.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 7.138094260134844e-08 [#---------] 8.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [#---------] 9.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [#---------] 10.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [#---------] 11.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [#---------] 12.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [#---------] 13.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [#---------] 14.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [##--------] 15.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [##--------] 16.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [##--------] 17.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [##--------] 18.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [##--------] 19.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [##--------] 20.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [##--------] 21.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [##--------] 22.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [##--------] 23.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [##--------] 24.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [##--------] 25.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [###-------] 26.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [###-------] 27.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [###-------] 28.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [###-------] 29.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [###-------] 30.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [###-------] 31.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [###-------] 32.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [###-------] 33.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [###-------] 34.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [####------] 35.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [####------] 36.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [####------] 37.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [####------] 38.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [####------] 39.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [####------] 40.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [####------] 41.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [####------] 42.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [####------] 43.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [####------] 44.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [####------] 45.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [#####-----] 46.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [#####-----] 47.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [#####-----] 48.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [#####-----] 49.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [#####-----] 50.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [#####-----] 51.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [#####-----] 52.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [#####-----] 53.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [#####-----] 54.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [######----] 55.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [######----] 56.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [######----] 57.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [######----] 58.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [######----] 59.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [######----] 60.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [######----] 61.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [######----] 62.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [######----] 63.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [######----] 64.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [######----] 65.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [#######---] 66.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [#######---] 67.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [#######---] 68.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [#######---] 69.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [#######---] 70.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [#######---] 71.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [#######---] 72.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [#######---] 73.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [#######---] 74.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [########--] 75.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [########--] 76.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [########--] 77.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [########--] 78.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [########--] 79.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [########--] 80.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [########--] 81.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [########--] 82.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [########--] 83.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [########--] 84.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [########--] 85.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [#########-] 86.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [#########-] 87.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [#########-] 88.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [#########-] 89.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [#########-] 90.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [#########-] 91.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [#########-] 92.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [#########-] 93.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [#########-] 94.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [##########] 95.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [##########] 96.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [##########] 97.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [##########] 98.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [##########] 99.00%
Anisotropic model: n_theta set to 1
spotpython tuning: 3.6762308162291255e-08 [##########] 100.00% Done...
Experiment saved to 01_res.pkl
min y: 3.6762308162291255e-08
x0: -0.00019173499462093834
[['x0', np.float64(-0.00019173499462093834)]]
Now we can start TensorBoard in the background. The TensorBoard process will read the TensorBoard files and visualize the hyperparameter tuning process. From the terminal, we can start TensorBoard with the following command:
tensorboard --logdir="./runs"
logdir
is the directory where the TensorBoard files are stored. In our case, the TensorBoard files are stored in the directory ./runs
.
TensorBoard will start a web server on port 6006. We can access the TensorBoard web server with the following URL:
http://localhost:6006/
The first TensorBoard visualization shows the objective function values plotted against the wall time. The wall time is the time that has passed since the start of the hyperparameter tuning process. The five initial design points are shown in the upper left region of the plot. The line visualizes the optimization process.
The second TensorBoard visualization shows the input values, i.e., \(x_0\), plotted against the wall time.
The third TensorBoard plot illustrates how spotpython
can be used as a microscope for the internal mechanisms of the surrogate-based optimization process. Here, one important parameter, the learning rate \(\theta\) of the Kriging surrogate is plotted against the number of optimization steps.
12.26 Jupyter Notebook
- The Jupyter-Notebook of this lecture is available on GitHub in the Hyperparameter-Tuning-Cookbook Repository