LightDataModule.setup(): stage: TrainerFn.FITTING
train_size: 0.36, val_size: 0.24 used for train & val data.
LightDataModule.val_dataloader(). Val. set size: 106
LightDataModule.train_dataloader(). data_train size: 160
LightDataModule.setup(): stage: TrainerFn.VALIDATING
LightDataModule.val_dataloader(). Val. set size: 106
train_model result: {'val_loss': 10240.3544921875, 'hp_metric': 10240.3544921875}
LightDataModule.setup(): stage: TrainerFn.FITTING
train_size: 0.36, val_size: 0.24 used for train & val data.
LightDataModule.val_dataloader(). Val. set size: 106
LightDataModule.train_dataloader(). data_train size: 160
LightDataModule.setup(): stage: TrainerFn.VALIDATING
LightDataModule.val_dataloader(). Val. set size: 106
train_model result: {'val_loss': nan, 'hp_metric': nan}
LightDataModule.setup(): stage: TrainerFn.FITTING
train_size: 0.36, val_size: 0.24 used for train & val data.
LightDataModule.val_dataloader(). Val. set size: 106
LightDataModule.train_dataloader(). data_train size: 160
LightDataModule.setup(): stage: TrainerFn.VALIDATING
LightDataModule.val_dataloader(). Val. set size: 106
train_model result: {'val_loss': nan, 'hp_metric': nan}
LightDataModule.setup(): stage: TrainerFn.FITTING
train_size: 0.36, val_size: 0.24 used for train & val data.
LightDataModule.val_dataloader(). Val. set size: 106
LightDataModule.train_dataloader(). data_train size: 160
LightDataModule.setup(): stage: TrainerFn.VALIDATING
LightDataModule.val_dataloader(). Val. set size: 106
train_model result: {'val_loss': 23857.75, 'hp_metric': 23857.75}
LightDataModule.setup(): stage: TrainerFn.FITTING
train_size: 0.36, val_size: 0.24 used for train & val data.
LightDataModule.val_dataloader(). Val. set size: 106
LightDataModule.train_dataloader(). data_train size: 160
LightDataModule.setup(): stage: TrainerFn.VALIDATING
LightDataModule.val_dataloader(). Val. set size: 106
train_model result: {'val_loss': nan, 'hp_metric': nan}
LightDataModule.setup(): stage: TrainerFn.FITTING
train_size: 0.36, val_size: 0.24 used for train & val data.
LightDataModule.val_dataloader(). Val. set size: 106
LightDataModule.train_dataloader(). data_train size: 160
LightDataModule.setup(): stage: TrainerFn.VALIDATING
LightDataModule.val_dataloader(). Val. set size: 106
train_model result: {'val_loss': 6874.3818359375, 'hp_metric': 6874.3818359375}
spotPython tuning: 6874.3818359375 [####------] 37.50%
LightDataModule.setup(): stage: TrainerFn.FITTING
train_size: 0.36, val_size: 0.24 used for train & val data.
LightDataModule.val_dataloader(). Val. set size: 106
LightDataModule.train_dataloader(). data_train size: 160
LightDataModule.setup(): stage: TrainerFn.VALIDATING
LightDataModule.val_dataloader(). Val. set size: 106
train_model result: {'val_loss': 7526.0234375, 'hp_metric': 7526.0234375}
spotPython tuning: 6874.3818359375 [#####-----] 50.00%
LightDataModule.setup(): stage: TrainerFn.FITTING
train_size: 0.36, val_size: 0.24 used for train & val data.
LightDataModule.val_dataloader(). Val. set size: 106
LightDataModule.train_dataloader(). data_train size: 160
LightDataModule.setup(): stage: TrainerFn.VALIDATING
LightDataModule.val_dataloader(). Val. set size: 106
train_model result: {'val_loss': 9060.6591796875, 'hp_metric': 9060.6591796875}
spotPython tuning: 6874.3818359375 [######----] 62.50%
LightDataModule.setup(): stage: TrainerFn.FITTING
train_size: 0.36, val_size: 0.24 used for train & val data.
LightDataModule.val_dataloader(). Val. set size: 106
LightDataModule.train_dataloader(). data_train size: 160
LightDataModule.setup(): stage: TrainerFn.VALIDATING
LightDataModule.val_dataloader(). Val. set size: 106
train_model result: {'val_loss': 9169.982421875, 'hp_metric': 9169.982421875}
spotPython tuning: 6874.3818359375 [########--] 75.00%
LightDataModule.setup(): stage: TrainerFn.FITTING
train_size: 0.36, val_size: 0.24 used for train & val data.
LightDataModule.val_dataloader(). Val. set size: 106
LightDataModule.train_dataloader(). data_train size: 160
LightDataModule.setup(): stage: TrainerFn.VALIDATING
LightDataModule.val_dataloader(). Val. set size: 106
train_model result: {'val_loss': 5747.65966796875, 'hp_metric': 5747.65966796875}
spotPython tuning: 5747.65966796875 [#########-] 87.50%
LightDataModule.setup(): stage: TrainerFn.FITTING
train_size: 0.36, val_size: 0.24 used for train & val data.
LightDataModule.val_dataloader(). Val. set size: 106
LightDataModule.train_dataloader(). data_train size: 160
LightDataModule.setup(): stage: TrainerFn.VALIDATING
LightDataModule.val_dataloader(). Val. set size: 106
train_model result: {'val_loss': 4623.73388671875, 'hp_metric': 4623.73388671875}
spotPython tuning: 4623.73388671875 [##########] 100.00% Done...
{'CHECKPOINT_PATH': 'runs/saved_models/',
'DATASET_PATH': 'data/',
'PREFIX': '037',
'RESULTS_PATH': 'results/',
'TENSORBOARD_PATH': 'runs/',
'_L_in': 10,
'_L_out': 1,
'_torchmetric': 'mean_squared_error',
'accelerator': 'auto',
'converters': None,
'core_model': <class 'spotPython.light.regression.netlightregression.NetLightRegression'>,
'core_model_hyper_dict': {'act_fn': {'class_name': 'spotPython.torch.activation',
'core_model_parameter_type': 'instance()',
'default': 'ReLU',
'levels': ['ReLU', 'LeakyReLU'],
'lower': 0,
'transform': 'None',
'type': 'factor',
'upper': 1},
'batch_size': {'default': 4,
'lower': 4,
'transform': 'transform_power_2_int',
'type': 'int',
'upper': 5},
'dropout_prob': {'default': 0.01,
'lower': 0.01,
'transform': 'None',
'type': 'float',
'upper': 0.1},
'epochs': {'default': 4,
'lower': 4,
'transform': 'transform_power_2_int',
'type': 'int',
'upper': 5},
'initialization': {'core_model_parameter_type': 'str',
'default': 'Default',
'levels': ['Default',
'Kaiming',
'Xavier'],
'lower': 0,
'transform': 'None',
'type': 'factor',
'upper': 2},
'l1': {'default': 3,
'lower': 3,
'transform': 'transform_power_2_int',
'type': 'int',
'upper': 8},
'lr_mult': {'default': 1.0,
'lower': 0.05,
'transform': 'None',
'type': 'float',
'upper': 1.0},
'optimizer': {'class_name': 'torch.optim',
'core_model_parameter_type': 'str',
'default': 'SGD',
'levels': ['Adam', 'RAdam'],
'lower': 0,
'transform': 'None',
'type': 'factor',
'upper': 1},
'patience': {'default': 2,
'lower': 2,
'transform': 'transform_power_2_int',
'type': 'int',
'upper': 3}},
'core_model_hyper_dict_default': {'act_fn': {'class_name': 'spotPython.torch.activation',
'core_model_parameter_type': 'instance()',
'default': 'ReLU',
'levels': ['Sigmoid',
'Tanh',
'ReLU',
'LeakyReLU',
'ELU',
'Swish'],
'lower': 0,
'transform': 'None',
'type': 'factor',
'upper': 5},
'batch_size': {'default': 4,
'lower': 1,
'transform': 'transform_power_2_int',
'type': 'int',
'upper': 4},
'dropout_prob': {'default': 0.01,
'lower': 0.0,
'transform': 'None',
'type': 'float',
'upper': 0.25},
'epochs': {'default': 4,
'lower': 4,
'transform': 'transform_power_2_int',
'type': 'int',
'upper': 9},
'initialization': {'core_model_parameter_type': 'str',
'default': 'Default',
'levels': ['Default',
'Kaiming',
'Xavier'],
'lower': 0,
'transform': 'None',
'type': 'factor',
'upper': 2},
'l1': {'default': 3,
'lower': 3,
'transform': 'transform_power_2_int',
'type': 'int',
'upper': 8},
'lr_mult': {'default': 1.0,
'lower': 0.1,
'transform': 'None',
'type': 'float',
'upper': 10.0},
'optimizer': {'class_name': 'torch.optim',
'core_model_parameter_type': 'str',
'default': 'SGD',
'levels': ['Adadelta',
'Adagrad',
'Adam',
'AdamW',
'SparseAdam',
'Adamax',
'ASGD',
'NAdam',
'RAdam',
'RMSprop',
'Rprop',
'SGD'],
'lower': 0,
'transform': 'None',
'type': 'factor',
'upper': 11},
'patience': {'default': 2,
'lower': 2,
'transform': 'transform_power_2_int',
'type': 'int',
'upper': 6}},
'core_model_name': None,
'counter': 8,
'data': None,
'data_dir': './data',
'data_module': None,
'data_set': <spotPython.data.diabetes.Diabetes object at 0x3c9a98f90>,
'data_set_name': None,
'db_dict_name': None,
'design': None,
'device': 'mps',
'devices': 1,
'enable_progress_bar': False,
'eval': None,
'fun_evals': 8,
'fun_repeats': 1,
'horizon': None,
'infill_criterion': 'y',
'k_folds': 3,
'log_graph': False,
'log_level': 50,
'loss_function': None,
'lower': array([3. , 4. , 1. , 0. , 0. , 0. , 0.1, 2. , 0. ]),
'max_surrogate_points': 30,
'max_time': 1,
'metric_params': {},
'metric_river': None,
'metric_sklearn': None,
'metric_sklearn_name': None,
'metric_torch': None,
'model_dict': {},
'n_points': 1,
'n_samples': None,
'n_total': None,
'noise': False,
'num_workers': 0,
'ocba_delta': 0,
'oml_grace_period': None,
'optimizer': None,
'path': None,
'prep_model': None,
'prep_model_name': None,
'progress_file': None,
'save_model': False,
'scenario': None,
'seed': 123,
'show_batch_interval': 1000000,
'show_models': False,
'show_progress': True,
'shuffle': None,
'sigma': 0.0,
'spot_tensorboard_path': 'runs/spot_logs/037_maans14_2024-04-22_02-15-20',
'spot_writer': <torch.utils.tensorboard.writer.SummaryWriter object at 0x3c9937a90>,
'target_column': None,
'target_type': None,
'task': None,
'test': None,
'test_seed': 1234,
'test_size': 0.4,
'tolerance_x': 1.4901161193847656e-08,
'train': None,
'upper': array([ 8. , 9. , 4. , 5. , 11. , 0.25, 10. , 6. , 2. ]),
'var_name': ['l1',
'epochs',
'batch_size',
'act_fn',
'optimizer',
'dropout_prob',
'lr_mult',
'patience',
'initialization'],
'var_type': ['int',
'int',
'int',
'factor',
'factor',
'float',
'float',
'int',
'factor'],
'verbosity': 0,
'weight_coeff': 0.0,
'weights': 1.0,
'weights_entry': None}