forked from oceansen/Erdre
-
Notifications
You must be signed in to change notification settings - Fork 3
/
Copy pathparams.yaml
129 lines (112 loc) · 2.89 KB
/
params.yaml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
# Broaching tool wear
profile:
dataset: broaching_with_tool_wear
clean:
target: tool_wear
classification: False
onehot_encode_target: False
combine_files: False
percentage_zeros_threshold: 1.0
correlation_metric: pearson
input_max_correlation_threshold: 1.0
featurize:
variables_to_include:
- acc1_x
- acc1_y
- acc1_z
- acc2_x
- acc2_y
- acc2_z
# By setting this to True, the add_-options below are overrided.
use_all_engineered_features_on_all_variables: True
# List the variables below each add_* to add engineered feature
add_sum:
# - variable1
add_gradient:
# - variable1
add_mean:
# - variable1
add_maximum:
# - variable1
add_minimum:
# - variable1
add_min_max_range:
# - variable1
add_slope:
# - variable1
add_slope_sin:
# - variable1
add_slope_cos:
# - variable1
add_standard_deviation:
# - variable1
add_variance:
# - variable1
add_peak_frequency:
# - variable1
rolling_window_size_sum: 128000
rolling_window_size_mean: 128000
rolling_window_size_max_min: 128000
rolling_window_size_standard_deviation: 128000
# List features here to remove the raw variables after engineering features
# from them
remove_features:
target_min_correlation_threshold: 0.0
split:
train_split: 0.6
shuffle_files: True
calibrate_split: 0.0
shuffle_samples_before_split: False
scale:
input: minmax
output:
sequentialize:
window_size: 1
overlap: 0
target_size: 1
shuffle_samples: True
future_predict: False
train:
seed: 2023
learning_method: gradientboosting
ensemble: True
hyperparameter_tuning: False
# Parameters for deep learning (dnn, cnn, rnn):
n_epochs: 1000
early_stopping: True
patience: 100
activation_function: relu
batch_size: 256
n_layers: 1
n_neurons: 8
dropout: 0.0
# Parameters for cnn and rnn
n_flattened_layers: 1
n_flattened_nodes: 16
# Parameters for cnn:
kernel_size: 10
maxpooling: True
maxpooling_size: 4
# Parameters for rnn:
unit_type: lstm
# Parameters for transformer:
ff_dim: 4
n_transformer_blocks: 4
n_heads: 4
head_size: 256
evaluate:
performance_metric: r2
threshold_for_ensemble_models: -1000
show_inputs: False
dropout_uncertainty_estimation: True
uncertainty_estimation_sampling_size: 1000
explain:
generate_explanations: True
number_of_background_samples: 100
number_of_summary_samples: 100
explanation_method: all
seed: 2020
combine_explanations:
combination_method: weighted
weighting_method: normalize
agreement_method: spearman