Installing and Starting

This section describes how to download and run Sparkling Water in different environments. Refer to the PySparkling and RSparkling sections for instructions on installing and running PySparkling and RSparkling.

Download and Run Locally

This section describes how to quickly get started with Sparkling Water on your personal computer (in Spark’s local cluster mode).

  1. Download and install Spark (if not already installed) from the Spark Downloads page.

    • Choose Spark release: 2.1.3

    • Choose package type: Pre-built for Hadoop 2.7 and later

  2. Point SPARK_HOME to the existing installation of Spark and export variable MASTER.

export SPARK_HOME="/path/to/spark/installation"
# To launch a local Spark cluster.
export MASTER="local[*]"
  1. From your terminal, run:

cd ~/Downloads
unzip sparkling-water-3.28.0.4-1-2.1.zip
cd sparkling-water-3.28.0.4-1-2.1
bin/sparkling-shell --conf "spark.executor.memory=1g"
  1. Create an H2O cloud inside the Spark cluster:

import org.apache.spark.h2o._
val h2oContext = H2OContext.getOrCreate(spark)
import h2oContext._
  1. Begin using Sparkling Water by following this demo, which imports airlines and weather data and runs predictions on delays.

Run on Hadoop

This section describes how to launch Sparkling Water on Hadoop using YARN.

  1. Download Spark (if not already installed) from the Spark Downloads page.

- Choose Spark release: 2.1.3
- Choose package type: Pre-built for Hadoop 2.7 and later
  1. Point SPARK_HOME to the existing installation of Spark.

export SPARK_HOME='/path/to/spark/installation'
  1. Set the HADOOP_CONF_DIR and Spark MASTER environmental variables.

export HADOOP_CONF_DIR=/etc/hadoop/conf
export MASTER="yarn"
  1. Download Spark and use sparkling-shell to launch Sparkling Shell on YARN.

wget http://h2o-release.s3.amazonaws.com/sparkling-water/spark-2.1/3.28.0.4-1-2.1/sparkling-water-3.28.0.4-1-2.1.zip
unzip sparkling-water-3.28.0.4-1-2.1.zip
cd sparkling-water-3.28.0.4-1-2.1/
bin/sparkling-shell --num-executors 3 --executor-memory 2g --master yarn --deploy-mode client
  1. Create an H2O cloud inside the Spark cluster:

import org.apache.spark.h2o._
val h2oContext = H2OContext.getOrCreate(spark)
import h2oContext._

Run on a Standalone Cluster

This section describes how to launch H2O on a standalone Spark cluster.

  1. Download Spark (if not already installed) from the Spark Downloads page.

- Choose Spark release: 2.1.3
- Choose package type: Pre-built for Hadoop 2.7 and later
  1. Point SPARK_HOME to the existing installation of Spark and export variable MASTER.

export SPARK_HOME='/path/to/spark/installation'
  1. From your terminal, run:

cd ~/Downloads
unzip sparkling-water-3.28.0.4-1-2.1.zip
cd sparkling-water-3.28.0.4-1-2.1
bin/launch-spark-cloud.sh
export MASTER="spark://localhost:7077"
bin/sparkling-shell
  1. Create an H2O cloud inside the Spark cluster:

import org.apache.spark.h2o._
val h2oContext = H2OContext.getOrCreate(spark)
import h2oContext._

Kluster Mode

Sparkling Water Kluster mode supports a connection to external H2O clusters (standalone/hadoop). The extended H2O cluster needs to be started with a corresponding H2O build, which can be downloaded below.

  1. Download and unpack the Sparkling Water distribution.

  2. Download the corresponding h2odriver for your Hadoop distribution (e.g., hdp2.2, cdh5.4) or standalone one:

bin/get-extended-h2o.sh standalone
  1. Start an H2O cluster, for example, in standalone mode:

java -cp h2odriver-extended.jar water.H2OApp -md5skip -name test
  1. In your Sparkling Water application, create H2OContext:

Scala

import org.apache.spark.h2o._
val conf = new H2OConf(spark).setExternalClusterMode().useManualClusterStart().setCloudName("test")
val hc = H2OContext.getOrCreate(spark, conf)

Python

from pysparkling import *
conf = H2OConf(spark).setExternalClusterMode().useManualClusterStart().setCloudName("test")
hc = H2OContext.getOrCreate(spark, conf)

Note: The following is a list of supported Hadoop distributions: SUBST_H2O_DRIVERS_LIST

For more information, please follow the Sparkling Water Backends.

Use from Maven

This section provides a gradle-style specification for Maven artifacts.

See the h2o-droplets GitHub repository for a working example.

repositories {
  mavenCentral()
}

dependencies {
  compile "ai.h2o:sparkling-water-package_2.11:3.28.0.4-1-2.1"
}

See Maven Central for artifact details.

Sparkling Water as a Spark Package

This section describes how to start Spark with Sparkling Water enabled via Spark package.

  1. Ensure that Spark is installed, and MASTER and SPARK_HOME environmental variables are properly set.

  2. Start Spark and point to maven coordinates of Sparkling Water:

$SPARK_HOME/bin/spark-shell --packages ai.h2o:sparkling-water-package_2.11:3.28.0.4-1-2.1
  1. Create an H2O cloud inside the Spark cluster:

import org.apache.spark.h2o._
val h2oContext = H2OContext.getOrCreate(spark)
import h2oContext._