H2O and Spark configuration
Convert given H2O frame into DataFrame type
Create a new H2OFrame based on existing Frame
Create a new H2OFrame based on existing Frame referenced by its key.
Transforms Dataset[Supported type] to H2OFrame
Transform DataFrame to H2OFrame
Transforms RDD[Supported type] to H2OFrame
A generic convert of Frame into Product RDD type
A generic convert of Frame into Product RDD type
Consider using asH2OFrame since asRDD has several limitations such as that asRDD can't be used in Spark REPL in case we are RDD[T] where T is class defined in REPL. This is because class T is created as inner class and we are not able to create instance of class T without outer scope - which is impossible to get.
This code: hc.asRDD[PUBDEV458Type](rdd) will need to be call as hc.asRDD[PUBDEV458Type].apply(rdd)
Convert given H2O frame into a Product RDD type
Convert given H2O frame into a Product RDD type
Consider using asH2OFrame since asRDD has several limitations such as that asRDD can't be used in Spark REPL in case we are RDD[T] where T is class defined in REPL. This is because class T is created as inner class and we are not able to create instance of class T without outer scope - which is impossible to get.
Used backend
directory where the logs will be downloaded
Return a copy of this H2OContext's configuration.
Return a copy of this H2OContext's configuration. The configuration cannot be changed at runtime.
Returns location of REST API of H2O client
Returns IP of H2O client
Returns port where H2O REST API is exposed
Runtime list of active H2O nodes
Define implicits available via h2oContext.implicits._
This method connects to external H2O cluster if spark.ext.h2o.externalClusterMode is set to true, otherwise it creates new H2O cluster living in Spark
Checks whether version of provided Spark is the same as Spark's version designated for this Sparkling Water version.
Checks whether version of provided Spark is the same as Spark's version designated for this Sparkling Water version. We check for correct version in shell scripts and during the build but we need to do the check also in the code in cases when the user executes for example spark-shell command with sparkling water assembly jar passed as --jars and initiates H2OContext. (Because in that case no check for correct Spark version has been done so far.)
Return true if running inside spark/sparkling water test.
Return true if running inside spark/sparkling water test.
Spark Context
true if the actual run is test run
Open H2O Flow running in this client.
Open browser for given address.
Open browser for given address.
address to open in browser, e.g., http://example.com
Set log level for the client running in driver
Set H2O log level for the client and all executors
Set log level for all H2O services running on executors
Spark Session
Stops H2O context.
Stops H2O context.
stop also spark context
Transforms Dataset[Supported type] to H2OFrame key
Transform DataFrame to H2OFrame key
Transforms RDD[Supported type] to H2OFrame key
Create new H2OContext based on provided H2O configuration