Deep Learning (Neural Networks)¶
H2O’s Deep Learning is based on a multi-layer feedforward artificial neural network that is trained with stochastic gradient descent using back-propagation. The network can contain a large number of hidden layers consisting of neurons with tanh, rectifier, and maxout activation functions. Advanced features such as adaptive learning rate, rate annealing, momentum training, dropout, L1 or L2 regularization, checkpointing, and grid search enable high predictive accuracy. Each compute node trains a copy of the global model parameters on its local data with multi-threading (asynchronously) and contributes periodically to the global model via model averaging across the network.
A feedforward artificial neural network (ANN) model, also known as deep neural network (DNN) or multi-layer perceptron (MLP), is the most common type of Deep Neural Network and the only type that is supported natively in H2O-3. Several other types of DNNs are popular as well, such as Convolutional Neural Networks (CNNs) and Recurrent Neural Networks (RNNs). MLPs work well on transactional (tabular) data; however if you have image data, then CNNs are a great choice. If you have sequential data (e.g. text, audio, time-series), then RNNs are a good choice. The H2O Deep Water project supports CNNs and RNNs though third-party integrations of other deep learning libraries such as TensorFlow, Caffe and MXNet. To learn more about how to use Deep Water, visit the H2O Deep Water repository and/or follow any of the available demos.
Quick Start and Additional Resources¶
- Deep Learning Booklet
- Deep Learning in H2O Tutorial (R): [GitHub]
- H2O + TensorFlow on AWS GPU Tutorial (Python Notebook) [Blog] [Github]
- Deep learning in H2O with Arno Candel (Overview) [Youtube]
- Top 10 tips and tricks [Youtube]
- NYC Tour Deep Learning Panel: Tensorflow, Mxnet, Caffe [Youtube]
- Deep Water project: [GitHub]
Defining a Deep Learning Model¶
H2O Deep Learning models have many input parameters, many of which are only accessible via the expert mode. For most cases, use the default values. Please read the following instructions before building extensive Deep Learning models. The application of grid search and successive continuation of winning models via checkpoint restart is highly recommended, as model performance can vary greatly.
model_id: (Optional) Specify a custom name for the model to use as a reference. By default, H2O automatically generates a destination key.
training_frame: (Required) Specify the dataset used to build the model. NOTE: In Flow, if you click the Build a model button from the
Parsecell, the training frame is entered automatically.
validation_frame: (Optional) Specify the dataset used to evaluate the accuracy of the model.
nfolds: Specify the number of folds for cross-validation.
Note: Cross-validation is not supported when autoencoder is enabled.
keep_cross_validation_predictions: Enable this option to keep the cross-validation predictions.
keep_cross_validation_fold_assignment: Enable this option to preserve the cross-validation fold assignment.
y: Specify the column to use as the dependent variable. The data can be numeric or categorical.
x: Specify a vector containing the names or indices of the predictor variables to use when building the model. If
xis missing, then all columns except
fold_assignment: (Applicable only if a value for nfolds is specified and fold_column is not specified) Specify the cross-validation fold assignment scheme. The available options are AUTO (which is Random), Random, Modulo, or Stratified (which will stratify the folds based on the response variable for classification problems).
fold_column: Specify the column that contains the cross-validation fold index assignment per observation.
ignored_columns: (Optional, Python and Flow only) Specify the column or columns to be excluded from the model. In Flow, click the checkbox next to a column name to add it to the list of columns excluded from the model. To add all columns, click the All button. To remove a column from the list of ignored columns, click the X next to the column name. To remove all columns from the list of ignored columns, click the None button. To search for a specific column, type the column name in the Search field above the column list. To only show columns with a specific percentage of missing values, specify the percentage in the Only show columns with more than 0% missing values field. To change the selections for the hidden columns, use the Select Visible or Deselect Visible buttons.
ignore_const_cols: Specify whether to ignore constant training columns, since no information can be gained from them. This option is enabled by default.
score_each_iteration: (Optional) Specify whether to score during each iteration of the model training.
weights_column: Specify a column to use for the observation weights, which are used for bias correction. The specified
weights_columnmust be included in the specified
Python only: To use a weights column when passing an H2OFrame to
xinstead of a list of column names, the specified
training_framemust contain the specified
Note: Weights are per-row observation weights. This is typically the number of times a row is repeated, but non-integer values are supported as well. During training, rows with higher weights matter more, due to the larger loss function pre-factor.
offset_column: (Applicable for regression only) Specify a column to use as the offset.
Note: Offsets are per-row “bias values” that are used during model training. For Gaussian distributions, they can be seen as simple corrections to the response (y) column. Instead of learning to predict the response (y-row), the model learns to predict the (row) offset of the response column. For other distributions, the offset corrections are applied in the linearized space before applying the inverse link function to get the actual response values. For more information, refer to the following link.
balance_classes: (Applicable for classification only) Specify whether to oversample the minority classes to balance the class distribution. This option is not enabled by default and can increase the data frame size. This option is only applicable for classification. Majority classes can be undersampled to satisfy the
class_sampling_factors: (Applicable only for classification and when
balance_classesis enabled) Specify the per-class (in lexicographical order) over/under-sampling ratios. By default, these ratios are automatically computed during training to obtain the class balance.
max_after_balance_size: Specify the maximum relative size of the training data after balancing class counts (balance_classes must be enabled). The value can be less than 1.0.
max_confusion_matrix_size: This option is deprecated and will be removed in a future release.
max_hit_ratio_k: Specify the maximum number (top K) of predictions to use for hit ratio computation. Applicable to multi-class only. To disable, enter 0.
checkpoint: Enter a model key associated with a previously-trained Deep Learning model. Use this option to build a new model as a continuation of a previously-generated model.
Note: Cross-validation is not supported during checkpoint restarts.
pretrained_autoencoder: Specify a pretrained autoencoder model to initialize this model with.
overwrite_with_best_model: Specify whether to overwrite the final model with the best model found during training, based on the option specified for stopping_metric. This option is enabled by default.
use_all_factor_levels: Specify whether to use all factor levels in the possible set of predictors; if you enable this option, sufficient regularization is required. By default, the first factor level is skipped. For Deep Learning models, this option is useful for determining variable importances and is automatically enabled if the autoencoder is selected.
standardize: If enabled, automatically standardize the data (mean 0, variance 1). If disabled, the user must provide properly scaled input data.
activation: Specify the activation function (Tahn, Tahn with dropout, Rectifier, Rectifier with dropout, Maxout, Maxout with dropout).
Note: Maxout is not supported when autoencoder is enabled.
hidden: Specify the hidden layer sizes (e.g., 100,100). The value must be positive.
epochs: Specify the number of times to iterate (stream) the dataset. The value can be a fraction.
train_samples_per_iteration: Specify the number of global training samples per MapReduce iteration. To specify one epoch, enter 0. To specify all available data (e.g., replicated training data), enter -1. To use the automatic values, enter -2.
target_ratio_comm_to_comp: Specify the target ratio of communication overhead to computation. This option is only enabled for multi-node operation and if train_samples_per_iteration equals -2 (auto-tuning).
seed: Specify the random number generator (RNG) seed for algorithm components dependent on randomization. The seed is consistent for each H2O instance so that you can create models with the same starting conditions in alternative configurations.
adaptive_rate: Specify whether to enable the adaptive learning rate (ADADELTA). This option is enabled by default.
rho: (Applicable only if adaptive_rate is enabled) Specify the adaptive learning rate time decay factor.
epsilon:(Applicable only if adaptive_rate is enabled) Specify the adaptive learning rate time smoothing factor to avoid dividing by zero.
rate: (Applicable only if adaptive_rate is disabled) Specify the learning rate. Higher values result in a less stable model, while lower values lead to slower convergence.
rate_annealing: (Applicable only if adaptive_rate is disabled) Specify the rate annealing value. The rate annealing is calculated as rate(1 + rate_annealing * samples).
rate_decay: (Applicable only if adaptive_rate is disabled) Specify the rate decay factor between layers. The rate decay is calculated as (N-th layer: rate * alpha^(N-1)).
momentum_start: (Applicable only if adaptive_rate is disabled) Specify the initial momentum at the beginning of training; we suggest 0.5.
momentum_ramp: (Applicable only if adaptive_rate is disabled) Specify the number of training samples for which the momentum increases.
momentum_stable: (Applicable only if adaptive_rate is disabled) Specify the final momentum after the ramp is over; we suggest 0.99.
nesterov_accelerated_gradient: (Applicable only if adaptive_rate is disabled) Enables the Nesterov Accelerated Gradient.
input_dropout_ratio: Specify the input layer dropout ratio to improve generalization. Suggested values are 0.1 or 0.2.
hidden_dropout_ratios: (Applicable only if the activation type is TanhWithDropout, RectifierWithDropout, or MaxoutWithDropout) Specify the hidden layer dropout ratio to improve generalization. Specify one value per hidden layer. The range is >= 0 to <1, and the default is 0.5.
l1: Specify the L1 regularization to add stability and improve generalization; sets the value of many weights to 0.
l2: Specify the L2 regularization to add stability and improve generalization; sets the value of many weights to smaller values.
max_w2: Specify the constraint for the squared sum of the incoming weights per unit (e.g., for Rectifier).
initial_weight_distribution: Specify the initial weight distribution (Uniform Adaptive, Uniform, or Normal).
initial_weight_scale: (Applicable only if initial_weight_distribution is Uniform or Normal) Specify the scale of the distribution function. For Uniform, the values are drawn uniformly. For Normal, the values are drawn from a Normal distribution with a standard deviation.
initial_weights: Specify a list of H2OFrame IDs to initialize the weight matrices of this model with.
initial_biases: Specify a list of H2OFrame IDs to initialize the bias vectors of this model with.
loss: Specify the loss function. The options are Automatic, CrossEntropy, Quadratic, Huber, or Absolute and the default value is Automatic.
- Use Absolute, Quadratic, or Huber for regression
- Use Absolute, Quadratic, Huber, or CrossEntropy for classification
distribution: Specify the distribution (i.e., the loss function). The options are AUTO, bernoulli, multinomial, gaussian, poisson, gamma, laplace, quantile, huber, or tweedie.
- If the distribution is
bernoulli, the the response column must be 2-class categorical
- If the distribution is
multinomial, the response column must be categorical.
- If the distribution is
poisson, the response column must be numeric.
- If the distribution is
laplace, the response column must be numeric.
- If the distribution is
tweedie, the response column must be numeric.
- If the distribution is
gaussian, the response column must be numeric.
- If the distribution is
huber, the response column must be numeric.
- If the distribution is
gamma, the response column must be numeric.
- If the distribution is
quantile, the response column must be numeric.
quantile_alpha: (Only applicable if
distribution="quantile".) Specify the quantile to be used for Quantile Regression.
tweedie_power: (Only applicable if
distribution="tweedie") Specify the Tweedie power. The range is from 1 to 2.
- For a normal distribution, enter
- For Poisson distribution, enter
- For a gamma distribution, enter
- For a compound Poisson-gamma distribution, enter a value greater than 1 but less than 2.
For more information, refer to Tweedie distribution.
- For a normal distribution, enter
huber_alpha: Specify the desired quantile for Huber/M-regression (the threshold between quadratic and linear loss). This value must be between 0 and 1.
score_interval: Specify the shortest time interval (in seconds) to wait between model scoring.
score_training_samples: Specify the number of training set samples for scoring. The value must be >= 0. To use all training samples, enter 0.
score_validation_samples: (Applicable only if a
validation_frameis specified) Specify the number of validation set samples for scoring. The value must be >= 0. To use all validation samples, enter 0.
score_duty_cycle: Specify the maximum duty cycle fraction forscoring. A lower value results in more training and a higher value results in more scoring.
classification_stop: This option specifies the stopping criteria in terms of classification error (1-accuracy) on the training data scoring dataset. When the error is at or below this threshold, training stops. To disable this option, enter -1.
regression_stop: (Regression models only) Specify the stopping criterion for regression error (MSE) on the training data. When the error is at or below this threshold, training stops. To disable this option, enter -1.
stopping_rounds: Stops training when the option selected for stopping_metric doesn’t improve for the specified number of training rounds, based on a simple moving average. To disable this feature, specify
0. The metric is computed on the validation data (if provided); otherwise, training data is used.
Note: If cross-validation is enabled:
- All cross-validation models stop training when the validation metric doesn’t improve.
- The main model runs for the mean number of epochs.
- N+1 models may be off by the number specified for stopping_rounds from the best model, but the cross-validation metric estimates the performance of the main model for the resulting number of epochs (which may be fewer than the specified number of epochs).
stopping_metric: Specify the metric to use for early stopping. The available options are:
auto: This defaults to
stopping_tolerance: Specify the relative tolerance for the metric-based stopping to stop training if the improvement is less than this value.
max_runtime_secs: Maximum allowed runtime in seconds for model training. Use 0 to disable.
score_validation_sampling: Specify the method used to sample validation dataset for scoring. This value can be either “Uniform” or “Stratified”.
diagnostics: Specify whether to compute the variable importances for input features (using the Gedeon method). For large networks, enabling this option can reduce speed. This option is enabled by default.
fast_mode: Specify whether to enable fast mode, a minor approximation in back-propagation. This option is enabled by default.
force_load_balance: Specify whether to force extra load balancing to increase training speed for small datasets and use all cores. This option is enabled by default.
variable_importances: Specify whether to compute variable importance. This option is not enabled by default.
replicate_training_data: Specify whether to replicate the entire training dataset onto every node for faster training on small datasets.
single_node_mode: Specify whether to run on a single node for fine-tuning of model parameters.
shuffle_training_data: Specify whether to shuffle the training data. This option is recommended if the training data is replicated and the value of train_samples_per_iteration is close to the number of nodes times the number of rows. This option is not enabled by default.
missing_values_handling: Specify how to handle missing values (Skip or MeanImputation).
quiet_mode: Specify whether to display less output in the standard output. This option is not enabled by default.
autoencoder: Specify whether to enable the Deep Learning autoencoder. This option is not enabled by default.
Note: Cross-validation is not supported when autoencoder is enabled.
sparse: Specify whether to enable sparse data handling, which is more efficient for data with many zero values.
col_major: Specify whether to use a column major weight matrix for the input layer. This option can speed up forward propagation but may reduce the speed of backpropagation. This option is not enabled by default.
average_activation: Specify the average activation for the sparse autoencoder. If Rectifier is used, the average_activation value must be positive.
sparsity_beta: (Applicable only if autoencoder is enabled) Specify the sparsity-based regularization optimization. For more information, refer to the following link.
max_categorical_features: Specify the maximum number of categorical features enforced via hashing. The value must be at least one.
reproducible: Specify whether to force reproducibility on small data. If this option is enabled, the model takes more time to generate because it uses only one thread.
export_weights_and_biases: Specify whether to export the neural network weights and biases as H2O frames.
mini_batch_size: Specify a value for the mini-batch size. (Smaller values lead to a better fit; larger values can speed up and generalize better.)
categorical_encoding: Specify one of the following encoding schemes for handling categorical features:
AUTO: Allow the algorithm to decide. In Deep Learning, the algorithm will perform
OneHotInternal: On the fly N+1 new cols for categorical features with N levels (default)
Binary: No more than 32 columns per categorical feature
Eigen: k columns per categorical feature, keeping projections of one-hot-encoded matrix onto k-dim eigen space only
LabelEncoder: Convert every enum into the integer of its index (for example, level 0 -> 0, level 1 -> 1, etc.). This is useful for keeping the number of columns small for XGBoost or DeepLearning/DeepWater, where the algorithm otherwise perform ExplicitOneHotEncoding.
SortByResponse: Reorders the levels by the mean response (for example, the level with lowest response -> 0, the level with second-lowest response -> 1, etc.). Note that this requires a specified response column.
Note: This value defaults to
one_hot_internal. Similarly, if
autois specified, then the algorithm performs
- elastic_averaging: Specify whether to enable elastic averaging between computing nodes, which can improve distributed model convergence.
- elastic_averaging_moving_rate: Specify the moving rate for elastic averaging. This option is only available if
- elastic_averaging_regularization: Specify the elastic averaging regularization strength. This option is only available if
- verbose: Print scoring history to the console. For Deep Learning, metrics are per epoch. This value defaults to FALSE.
Interpreting a Deep Learning Model¶
To view the results, click the View button. The output for the Deep Learning model includes the following information for both the training and testing sets:
- Model parameters (hidden)
- A chart of the variable importances
- A graph of the scoring history (training MSE and validation MSE vs epochs)
- Training and validation metrics confusion matrix
- Output (model category, weights, biases)
- Status of neuron layers (layer number, units, type, dropout, L1, L2, mean rate, rate RMS, momentum, mean weight, weight RMS, mean bias, bias RMS)
- Scoring history in tabular format
- Training and validation metrics (model name, model checksum name, frame name, frame checksum name, description, model category, duration in ms, scoring time, predictions, MSE, R2, logloss)
- Top-K Hit Ratios for training and validation (for multi-class classification)
- How does the algorithm handle missing values during training?
Depending on the selected missing value handling policy, they are either imputed mean or the whole row is skipped. The default behavior is mean imputation. Note that categorical variables are imputed by adding an extra “missing” level. Optionally, Deep Learning can skip all rows with any missing values.
- How does the algorithm handle missing values during testing?
Missing values in the test set will be mean-imputed during scoring.
- What happens if the response has missing values?
No errors will occur, but nothing will be learned from rows containing missing the response.
- What happens when you try to predict on a categorical level not seen during training?
For an unseen categorical level in the test set, Deep Learning makes an extra input neuron that remains untrained and contributes some random amount to the subsequent layer.
- Does it matter if the data is sorted?
Yes, since the training set is processed in order. Depending whether
train_samples_per_iterationis enabled, some rows will be skipped. If
shuffle_training_datais enabled, then each thread that is processing a small subset of rows will process rows randomly, but it is not a global shuffle.
- Should data be shuffled before training?
Yes, the data should be shuffled before training, especially if the dataset is sorted.
- How does the algorithm handle highly imbalanced data in a response column?
max_after_balance_sizeto control over/under-sampling.
- What if there are a large number of columns?
The input neuron layer’s size is scaled to the number of input features, so as the number of columns increases, the model complexity increases as well.
- What if there are a large number of categorical factor levels?
This is something to look out for. Say you have three columns: zip code (70k levels), height, and income. The resulting number of internally one-hot encoded features will be 70,002 and only 3 of them will be activated (non-zero). If the first hidden layer has 200 neurons, then the resulting weight matrix will be of size 70,002 x 200, which can take a long time to train and converge. In this case, we recommend either reducing the number of categorical factor levels upfront (e.g., using
h2o.interaction()from R), or specifying
max_categorical_featuresto use feature hashing to reduce the dimensionality.
- How does your Deep Learning Autoencoder work? Is it deep or shallow?
H2O’s DL autoencoder is based on the standard deep (multi-layer) neural net architecture, where the entire network is learned together, instead of being stacked layer-by-layer. The only difference is that no response is required in the input and that the output layer has as many neurons as the input layer. If you don’t achieve convergence, then try using the Tanh activation and fewer layers. We have some example test scripts here, and even some that show how stacked auto-encoders can be implemented in R.
- When building the model, does Deep Learning use all features or a selection of the best features?
For Deep Learning, all features are used, unless you manually specify that columns should be ignored. Adding an L1 penalty can make the model sparse, but it is still the full size.
- What is the relationship between iterations, epochs, and the ``train_samples_per_iteration`` parameter?
Epochs measures the amount of training. An iteration is one MapReduce (MR) step - essentially, one pass over the data. The
train_samples_per_iterationparameter is the amount of data to use for training for each MR step, which can be more or less than the number of rows.
- When do ``reduce()`` calls occur, after each iteration or each epoch?
reduce()calls occur after every two
map()calls, between threads and ultimately between nodes. There are many
reduce()calls, much more than one per MapReduce step (also known as an “iteration”). Epochs are not related to MR iterations, unless you specify
-1(or to number of rows/nodes). Otherwise, one MR iteration can train with an arbitrary number of training samples (as specified by
- Does each Mapper task work on a separate neural-net model that is combined during reduction, or is each Mapper manipulating a shared object that’s persistent across nodes?
Neither; there’s one model per compute node, so multiple Mappers/threads share one model, which is why H2O is not reproducible unless a small dataset is used and
reproducible=T, which effectively rebalances to a single chunk and leads to only one thread to launch a
map(). The current behavior is simple model averaging; between-node model averaging via “Elastic Averaging” is currently in progress.
- Is the loss function and backpropagation performed after each individual training sample, each iteration, or at the epoch level?
Loss function and backpropagation are performed after each training sample (mini-batch size 1 == online stochastic gradient descent).
- When using Hinton’s dropout and specifying an input dropout ratio of ~20% and ``train_samples_per_iteration`` is set to 50, will each of the 50 samples have a different set of the 20% input neurons suppressed?
Yes - suppression is not done at the iteration level across as samples in that iteration. The dropout mask is different for each training sample.
- When using dropout parameters such as ``input_dropout_ratio``, what happens if you use only ``Rectifier`` instead of ``RectifierWithDropout`` in the activation parameter?
The amount of dropout on the input layer can be specified for all activation functions, but hidden layer dropout is only supported is set to
WithDropout. The default hidden dropout is 50%, so you don’t need to specify anything but the activation type to get good results, but you can set the hidden dropout values for each layer separately.
- When using the ``score_validation_sampling`` and ``score_training_samples`` parameters, is scoring done at the end of the Deep Learning run?
The majority of scoring takes place after each MR iteration. After the iteration is complete, it may or may not be scored, depending on two criteria: the time since the last scoring and the time needed for scoring.
The maximum time between scoring (
score_interval, default = 5 seconds) and the maximum fraction of time spent scoring (
score_duty_cycle) independently of loss function, backpropagation, etc.
Of course, using more training or validation samples will increase the time for scoring, as well as scoring more frequently. For more information about how this affects runtime, refer to the Deep Learning Performance Guide.
- How does the validation frame affect the built neuron network?
The validation frame is only used for scoring and does not directly affect the model. However, the validation frame can be used stopping the model early if
overwrite_with_best_model = T, which is the default. If this parameter is enabled, the model with the lowest validation error is displayed at the end of the training.
By default, the validation frame is used to tune the model parameters (such as number of epochs) and will return the best model as measured by the validation metrics, depending on how often the validation metrics are computed (
score_duty_cycle) and whether the validation frame itself was sampled.
Model-internal sampling of the validation frame (
score_validation_samplingfor optional stratification) will affect early stopping quality. If you specify a validation frame but set
score_validation_samplesto more than the number of rows in the validation frame (instead of 0, which represents the entire frame), the validation metrics received at the end of training will not be reproducible, since the model does internal sampling.
- Are there any best practices for building a model using checkpointing?
In general, to get the best possible model, we recommend building a model with
train_samples_per_iteration = -2(which is the default value for auto-tuning) and saving it.
To improve the initial model, start from the previous model and add iterations by building another model, setting the checkpoint to the previous model, and changing
target_ratio_comm_to_comp, or other parameters.
If you don’t know your model ID because it was generated by R, look it up using
h2o.ls(). By default, Deep Learning model names start with
deeplearning_To view the model, use
m <- h2o.getModel("my_model_id")or
There are a few ways to manage checkpoint restarts:
Option 1: (Multi-node only) Leave
train_samples_per_iteration = -2, increase
target_comm_to_compfrom 0.05 to 0.25 or 0.5, which provides more communication. This should result in a better model when using multiple nodes. Note: This does not affect single-node performance.
Option 2: (Single or multi-node) Set
train_samples_per_iterationto (N), where (N) is the number of training samples used for training by the entire cluster for one iteration. Each of the nodes then trains on (N) randomly-chosen rows for every iteration. The number defined as (N) depends on the dataset size and the model complexity.
Option 3: (Single or multi-node) Change regularization parameters such as
l1, l2, max_w2, input_droput_ratioor
hidden_dropout_ratios. We recommend build the first mode using
input_dropout_ratio = 0(if there is suspected noise in the input), and
hidden_dropout_ratios=c(0,0,0)(for the ability to enable dropout regularization later).
- How does class balancing work?
max_after_balance_sizeparameter defines the maximum size of the over-sampled dataset. For example, if
max_after_balance_size = 3, the over-sampled dataset will not be greater than three times the size of the original dataset.
For example, if you have five classes with priors of 90%, 2.5%, 2.5%, and 2.5% (out of a total of one million rows) and you oversample to obtain a class balance using
balance_classes = T, the result is all four minor classes are oversampled by forty times and the total dataset will be 4.5 times as large as the original dataset (900,000 rows of each class). If
max_after_balance_size = 3, all five balance classes are reduced by 3/5 resulting in 600,000 rows each (three million total).
To specify the per-class over- or under-sampling factors, use
class_sampling_factors. In the previous example, the default behavior with
balance_classesis equivalent to
c(1,40,40,40,40), while when
max_after_balance\size = 3, the results would be
In all cases, the probabilities are adjusted to the pre-sampled space, so the minority classes will have lower average final probabilities than the majority class, even if they were sampled to reach class balance.
- How is variable importance calculated for Deep Learning?
For Deep Learning, variable importance is calculated using the Gedeon method.
- How is deviance computed for a Deep Learning regression model?
The following formula is used to compute deviance for a Deep Learning regression model:Loss = Quadratic -> MSE==Deviance For Absolute/Laplace or Huber -> MSE != Deviance
Niu, Feng, et al. “Hogwild!: A lock-free approach to parallelizing stochastic gradient descent.” Advances in Neural Information Processing Systems 24 (2011): 693-701. (algorithm implemented is on p.5)