y
¶
Available in: GBM, DRF, Deep Learning, GLM, GAM, HGLM, Naïve-Bayes, Stacked Ensembles, AutoML, XGBoost, Aggregator, Uplift DRF, AdaBoost, Decision Tree, ANOVAGLM, ModelSelection
Hyperparameter: no
Description¶
Use this option to specify a response column (y-axis). The response column is the column that you are attempting to predict. For example, based on a set of parameters in a training dataset, will a new customber be more or less likely to purchase a product? Or based on some known variables, what is the likelihood that a flight will be delayed? In both cases, a model can be applied to a training frame and to a validation frame to predict the likely response.
Response Columns with DL and GBM Distribution
Response columns can be numeric or categorical, and they can be binomial or multiomial. If you are specifying a distribution type in DL or GBM, however, then keep in mind the following when defining a response column:
If the distribution is
bernoulli
, the the response column must be 2-class categoricalIf the distribution is
multinomial
, the response column must be categorical.If the distribution is
poisson
, the response column must be numeric.If the distribution is
laplace
, the response column must be numeric.If the distribution is
tweedie
, the response column must be numeric.If the distribution is
gaussian
, the response column must be numeric.If the distribution is
huber
, the response column must be numeric.If the distribution is
gamma
, the response column must be numeric.If the distribution is
quantile
, the response column must be numeric.
Response Columns with GLM Family
In GLM, you can specify one of the following family options based on the response column type:
gaussian
: The data must be numeric (Real or Int). This is the default family.binomial
: The data must be categorical 2 levels/classes or binary (Enum or Int).quasibinomial
: The data must be numeric.multinomial
: The data can be categorical with more than two levels/classes (Enum).poisson
: The data must be numeric and non-negative (Int).gamma
: The data must be numeric and continuous and positive (Real or Int).tweedie
: The data must be numeric and continuous (Real) and non-negative.ordinal
: Requires a categorical response with at least 3 levels. (For 2-class problems use family=”binomial”.)
Notes:
The response column cannot be the same as the fold_column.
For supervised learning, the response column cannot be the same as the weights_column, and the response column must exist in both the training frame and in the validation frame.
Example¶
library(h2o)
h2o.init()
# import the cars dataset:
# this dataset is used to classify whether or not a car is economical based on
# the car's displacement, power, weight, and acceleration, and the year it was made
cars <- h2o.importFile("https://s3.amazonaws.com/h2o-public-test-data/smalldata/junit/cars_20mpg.csv")
# convert response column to a factor
cars["economy_20mpg"] <- as.factor(cars["economy_20mpg"])
# set the predictor names and the response column name
predictors <- c("displacement", "power", "weight", "acceleration", "year")
response <- "economy_20mpg"
# split into train and validation sets
cars_split <- h2o.splitFrame(data = cars, ratios = 0.8, seed = 1234)
train <- cars_split[[1]]
valid <- cars_split[[2]]
# try using the `y` parameter:
# train your model, where you specify your 'x' predictors, your 'y' the response column
# training_frame and validation_frame
cars_gbm <- h2o.gbm(x = predictors, y = response, training_frame = train,
validation_frame = valid, seed = 1234)
# print the auc for your model
print(h2o.auc(cars_gbm, valid = TRUE))
import h2o
from h2o.estimators.gbm import H2OGradientBoostingEstimator
h2o.init()
h2o.cluster().show_status()
# import the cars dataset:
# this dataset is used to classify whether or not a car is economical based on
# the car's displacement, power, weight, and acceleration, and the year it was made
cars = h2o.import_file("https://s3.amazonaws.com/h2o-public-test-data/smalldata/junit/cars_20mpg.csv")
# convert response column to a factor
cars["economy_20mpg"] = cars["economy_20mpg"].asfactor()
# set the predictor names and the response column name
predictors = ["displacement","power","weight","acceleration","year"]
response = "economy_20mpg"
# split into train and validation sets
train, valid = cars.split_frame(ratios = [.8], seed = 1234)
# try using the `y` parameter:
# first initialize your estimator
cars_gbm = H2OGradientBoostingEstimator(seed = 1234)
# then train your model, where you specify your 'x' predictors, your 'y' the response column
# training_frame and validation_frame
cars_gbm.train(x = predictors, y = response, training_frame = train, validation_frame = valid)
# print the auc for the validation data
cars_gbm.auc(valid=True)