Panel A: ENet . | Panel B: RF . |
---|---|
Package: | Package: |
Scikit-learn (SGDRegressor) | Scikit-learn (RandomForestRegressor) |
Feature transformation: | Feature transformation: |
Standard & robust scaling | Standard & robust scaling |
Selection by variance threshold | Selection by variance threshold |
Model parameters: | Model parameters: |
L1-L2-penalty: | Number of trees: 300 |
L1-ratio: | Max. depth: |
Max. features: | |
Optimization: | |
Stochastic gradient descent | |
Tolerance: 10−4 | |
Max. epochs: 1,000 | |
Learning rate: | |
Random search: | Random search: |
Number of combinations: 1,000 | Number of combinations: 500 |
Panel A: ENet . | Panel B: RF . |
---|---|
Package: | Package: |
Scikit-learn (SGDRegressor) | Scikit-learn (RandomForestRegressor) |
Feature transformation: | Feature transformation: |
Standard & robust scaling | Standard & robust scaling |
Selection by variance threshold | Selection by variance threshold |
Model parameters: | Model parameters: |
L1-L2-penalty: | Number of trees: 300 |
L1-ratio: | Max. depth: |
Max. features: | |
Optimization: | |
Stochastic gradient descent | |
Tolerance: 10−4 | |
Max. epochs: 1,000 | |
Learning rate: | |
Random search: | Random search: |
Number of combinations: 1,000 | Number of combinations: 500 |
Panel C: GBRT . | Panel D: ANN . |
---|---|
Package: | Package: |
Scikit-learn (GradientBoostingRegressor) | Tensorflow/Keras (Sequential) |
Feature transformation: | Feature transformation: |
Standard & robust scaling | Standard & robust scaling |
Selection by variance threshold | Selection by variance threshold |
Model parameters: | Model parameters: |
Number of trees: | Activation: TanH (Glorot), ReLU (He) |
Max. depth: | Hidden layers: |
Max. features: | First hidden layer nodes: |
Learning rate: | Network architecture: Pyramid |
Max. weight norm: 4 | |
Dropout rate: | |
L1-penalty: | |
Optimization: | |
Adaptive moment estimation | |
Batch size: | |
Learning rate: | |
Early stopping patience: 6 | |
Max. epochs: 50 | |
Batch normalization before activ. | |
Number of networks in ensemble: 10 | |
Random search: | Random search: |
Number of combinations: 300 | Number of combinations: 1,000 |
Panel C: GBRT . | Panel D: ANN . |
---|---|
Package: | Package: |
Scikit-learn (GradientBoostingRegressor) | Tensorflow/Keras (Sequential) |
Feature transformation: | Feature transformation: |
Standard & robust scaling | Standard & robust scaling |
Selection by variance threshold | Selection by variance threshold |
Model parameters: | Model parameters: |
Number of trees: | Activation: TanH (Glorot), ReLU (He) |
Max. depth: | Hidden layers: |
Max. features: | First hidden layer nodes: |
Learning rate: | Network architecture: Pyramid |
Max. weight norm: 4 | |
Dropout rate: | |
L1-penalty: | |
Optimization: | |
Adaptive moment estimation | |
Batch size: | |
Learning rate: | |
Early stopping patience: 6 | |
Max. epochs: 50 | |
Batch normalization before activ. | |
Number of networks in ensemble: 10 | |
Random search: | Random search: |
Number of combinations: 300 | Number of combinations: 1,000 |
Notes: This table shows the hyperparameter search space and the Python packages used for both long and short training. Parameter configurations not listed here correspond to the respective default settings.
Panel A: ENet . | Panel B: RF . |
---|---|
Package: | Package: |
Scikit-learn (SGDRegressor) | Scikit-learn (RandomForestRegressor) |
Feature transformation: | Feature transformation: |
Standard & robust scaling | Standard & robust scaling |
Selection by variance threshold | Selection by variance threshold |
Model parameters: | Model parameters: |
L1-L2-penalty: | Number of trees: 300 |
L1-ratio: | Max. depth: |
Max. features: | |
Optimization: | |
Stochastic gradient descent | |
Tolerance: 10−4 | |
Max. epochs: 1,000 | |
Learning rate: | |
Random search: | Random search: |
Number of combinations: 1,000 | Number of combinations: 500 |
Panel A: ENet . | Panel B: RF . |
---|---|
Package: | Package: |
Scikit-learn (SGDRegressor) | Scikit-learn (RandomForestRegressor) |
Feature transformation: | Feature transformation: |
Standard & robust scaling | Standard & robust scaling |
Selection by variance threshold | Selection by variance threshold |
Model parameters: | Model parameters: |
L1-L2-penalty: | Number of trees: 300 |
L1-ratio: | Max. depth: |
Max. features: | |
Optimization: | |
Stochastic gradient descent | |
Tolerance: 10−4 | |
Max. epochs: 1,000 | |
Learning rate: | |
Random search: | Random search: |
Number of combinations: 1,000 | Number of combinations: 500 |
Panel C: GBRT . | Panel D: ANN . |
---|---|
Package: | Package: |
Scikit-learn (GradientBoostingRegressor) | Tensorflow/Keras (Sequential) |
Feature transformation: | Feature transformation: |
Standard & robust scaling | Standard & robust scaling |
Selection by variance threshold | Selection by variance threshold |
Model parameters: | Model parameters: |
Number of trees: | Activation: TanH (Glorot), ReLU (He) |
Max. depth: | Hidden layers: |
Max. features: | First hidden layer nodes: |
Learning rate: | Network architecture: Pyramid |
Max. weight norm: 4 | |
Dropout rate: | |
L1-penalty: | |
Optimization: | |
Adaptive moment estimation | |
Batch size: | |
Learning rate: | |
Early stopping patience: 6 | |
Max. epochs: 50 | |
Batch normalization before activ. | |
Number of networks in ensemble: 10 | |
Random search: | Random search: |
Number of combinations: 300 | Number of combinations: 1,000 |
Panel C: GBRT . | Panel D: ANN . |
---|---|
Package: | Package: |
Scikit-learn (GradientBoostingRegressor) | Tensorflow/Keras (Sequential) |
Feature transformation: | Feature transformation: |
Standard & robust scaling | Standard & robust scaling |
Selection by variance threshold | Selection by variance threshold |
Model parameters: | Model parameters: |
Number of trees: | Activation: TanH (Glorot), ReLU (He) |
Max. depth: | Hidden layers: |
Max. features: | First hidden layer nodes: |
Learning rate: | Network architecture: Pyramid |
Max. weight norm: 4 | |
Dropout rate: | |
L1-penalty: | |
Optimization: | |
Adaptive moment estimation | |
Batch size: | |
Learning rate: | |
Early stopping patience: 6 | |
Max. epochs: 50 | |
Batch normalization before activ. | |
Number of networks in ensemble: 10 | |
Random search: | Random search: |
Number of combinations: 300 | Number of combinations: 1,000 |
Notes: This table shows the hyperparameter search space and the Python packages used for both long and short training. Parameter configurations not listed here correspond to the respective default settings.
This PDF is available to Subscribers Only
View Article Abstract & Purchase OptionsFor full access to this pdf, sign in to an existing account, or purchase an annual subscription.